You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by jc...@apache.org on 2018/05/25 00:57:55 UTC
[23/58] [abbrv] [partial] hive git commit: HIVE-19617: Rename test
tables to avoid collisions during execution in batches (Jesus Camacho
Rodriguez, reviewed by Gunther Hagleitner)
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/multi_insert_move_tasks_share_dependencies.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_insert_move_tasks_share_dependencies.q b/ql/src/test/queries/clientpositive/multi_insert_move_tasks_share_dependencies.q
index 3ddaa47..cd6073b 100644
--- a/ql/src/test/queries/clientpositive/multi_insert_move_tasks_share_dependencies.q
+++ b/ql/src/test/queries/clientpositive/multi_insert_move_tasks_share_dependencies.q
@@ -2,23 +2,23 @@ set hive.multi.insert.move.tasks.share.dependencies=true;
set hive.stats.dbclass=fs;
-- SORT_QUERY_RESULTS
-create table src_multi1 like src;
-create table src_multi2 like src;
+create table src_multi1_n4 like src;
+create table src_multi2_n5 like src;
set hive.merge.mapfiles=false;
set hive.merge.mapredfiles=false;
explain
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20;
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n4;
+select * from src_multi2_n5;
set hive.merge.mapfiles=true;
@@ -26,45 +26,45 @@ set hive.merge.mapredfiles=false;
explain
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20;
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n4;
+select * from src_multi2_n5;
set hive.merge.mapfiles=false;
set hive.merge.mapredfiles=true;
explain
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20;
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n4;
+select * from src_multi2_n5;
set hive.merge.mapfiles=true;
set hive.merge.mapredfiles=true;
explain
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20;
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n4;
+select * from src_multi2_n5;
@@ -73,15 +73,15 @@ set hive.merge.mapredfiles=false;
explain
from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n4 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20 group by key, value;
from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n4 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20 group by key, value;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n4;
+select * from src_multi2_n5;
set hive.merge.mapfiles=false;
@@ -89,30 +89,30 @@ set hive.merge.mapredfiles=true;
explain
from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n4 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20 group by key, value;
from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n4 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20 group by key, value;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n4;
+select * from src_multi2_n5;
set hive.merge.mapfiles=true;
set hive.merge.mapredfiles=false;
explain
from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n4 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20 group by key, value;
from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n4 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20 group by key, value;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n4;
+select * from src_multi2_n5;
set hive.merge.mapfiles=true;
@@ -120,15 +120,15 @@ set hive.merge.mapredfiles=true;
explain
from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n4 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20 group by key, value;
from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n4 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20 group by key, value;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n4;
+select * from src_multi2_n5;
@@ -138,60 +138,60 @@ set hive.merge.mapredfiles=false;
explain
from (select * from src union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20;
from (select * from src union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n4;
+select * from src_multi2_n5;
set hive.merge.mapfiles=true;
set hive.merge.mapredfiles=false;
explain
from (select * from src union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20;
from (select * from src union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n4;
+select * from src_multi2_n5;
set hive.merge.mapfiles=false;
set hive.merge.mapredfiles=true;
explain
from (select * from src union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20;
from (select * from src union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n4;
+select * from src_multi2_n5;
set hive.merge.mapfiles=true;
set hive.merge.mapredfiles=true;
explain
from (select * from src union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20;
from (select * from src union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n4;
+select * from src_multi2_n5;
@@ -332,19 +332,19 @@ set hive.merge.mapredfiles=false;
explain
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20
insert overwrite local directory '${system:test.tmp.dir}/hive_test/multiins_local/1' select * where key < 10 group by key, value cluster by key
insert overwrite local directory '${system:test.tmp.dir}/hive_test/multiins_local/2' select * where key > 10 and key < 20 group by key, value cluster by value;
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20
insert overwrite local directory '${system:test.tmp.dir}/hive_test/multiins_local/1' select * where key < 10 group by key, value cluster by key
insert overwrite local directory '${system:test.tmp.dir}/hive_test/multiins_local/2' select * where key > 10 and key < 20 group by key, value cluster by value;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n4;
+select * from src_multi2_n5;
dfs -ls ${system:test.tmp.dir}/hive_test/multiins_local;
dfs -rmr ${system:test.tmp.dir}/hive_test/multiins_local;
@@ -354,19 +354,19 @@ set hive.merge.mapredfiles=true;
explain
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20
insert overwrite local directory '${system:test.tmp.dir}/hive_test/multiins_local/1' select * where key < 10 group by key, value cluster by key
insert overwrite local directory '${system:test.tmp.dir}/hive_test/multiins_local/2' select * where key > 10 and key < 20 group by key, value cluster by value;
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20
insert overwrite local directory '${system:test.tmp.dir}/hive_test/multiins_local/1' select * where key < 10 group by key, value cluster by key
insert overwrite local directory '${system:test.tmp.dir}/hive_test/multiins_local/2' select * where key > 10 and key < 20 group by key, value cluster by value;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n4;
+select * from src_multi2_n5;
dfs -ls ${system:test.tmp.dir}/hive_test/multiins_local;
dfs -rmr ${system:test.tmp.dir}/hive_test/multiins_local;
@@ -376,19 +376,19 @@ set hive.merge.mapredfiles=false;
explain
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20
insert overwrite local directory '${system:test.tmp.dir}/hive_test/multiins_local/1' select * where key < 10 group by key, value cluster by key
insert overwrite local directory '${system:test.tmp.dir}/hive_test/multiins_local/2' select * where key > 10 and key < 20 group by key, value cluster by value;
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20
insert overwrite local directory '${system:test.tmp.dir}/hive_test/multiins_local/1' select * where key < 10 group by key, value cluster by key
insert overwrite local directory '${system:test.tmp.dir}/hive_test/multiins_local/2' select * where key > 10 and key < 20 group by key, value cluster by value;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n4;
+select * from src_multi2_n5;
dfs -ls ${system:test.tmp.dir}/hive_test/multiins_local;
dfs -rmr ${system:test.tmp.dir}/hive_test/multiins_local;
@@ -398,19 +398,19 @@ set hive.merge.mapredfiles=true;
explain
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20
insert overwrite local directory '${system:test.tmp.dir}/hive_test/multiins_local/1' select * where key < 10 group by key, value cluster by key
insert overwrite local directory '${system:test.tmp.dir}/hive_test/multiins_local/2' select * where key > 10 and key < 20 group by key, value cluster by value;
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20
+insert overwrite table src_multi1_n4 select * where key < 10
+insert overwrite table src_multi2_n5 select * where key > 10 and key < 20
insert overwrite local directory '${system:test.tmp.dir}/hive_test/multiins_local/1' select * where key < 10 group by key, value cluster by key
insert overwrite local directory '${system:test.tmp.dir}/hive_test/multiins_local/2' select * where key > 10 and key < 20 group by key, value cluster by value;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n4;
+select * from src_multi2_n5;
dfs -ls ${system:test.tmp.dir}/hive_test/multiins_local;
dfs -rmr ${system:test.tmp.dir}/hive_test/multiins_local;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/multi_insert_partitioned.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_insert_partitioned.q b/ql/src/test/queries/clientpositive/multi_insert_partitioned.q
index 0d78c72..d818613 100644
--- a/ql/src/test/queries/clientpositive/multi_insert_partitioned.q
+++ b/ql/src/test/queries/clientpositive/multi_insert_partitioned.q
@@ -4,46 +4,46 @@ set hive.mapred.mode=nonstrict;
set hive.explain.user=false;
set hive.exec.dynamic.partition.mode=nonstrict;
-drop table intermediate;
+drop table intermediate_n3;
-create table intermediate(key int) partitioned by (p int) stored as orc;
-insert into table intermediate partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2;
-insert into table intermediate partition(p='456') select distinct key from src where key is not null order by key asc limit 2;
-insert into table intermediate partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2;
+create table intermediate_n3(key int) partitioned by (p int) stored as orc;
+insert into table intermediate_n3 partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2;
+insert into table intermediate_n3 partition(p='456') select distinct key from src where key is not null order by key asc limit 2;
+insert into table intermediate_n3 partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2;
drop table multi_partitioned;
create table multi_partitioned (key int, key2 int) partitioned by (p int);
-from intermediate
+from intermediate_n3
insert into table multi_partitioned partition(p=1) select p, key
insert into table multi_partitioned partition(p=2) select key, p;
select * from multi_partitioned order by key, key2, p;
desc formatted multi_partitioned;
-from intermediate
+from intermediate_n3
insert overwrite table multi_partitioned partition(p=2) select p, key
insert overwrite table multi_partitioned partition(p=1) select key, p;
select * from multi_partitioned order by key, key2, p;
desc formatted multi_partitioned;
-from intermediate
+from intermediate_n3
insert into table multi_partitioned partition(p=2) select p, key
insert overwrite table multi_partitioned partition(p=1) select key, p;
select * from multi_partitioned order by key, key2, p;
desc formatted multi_partitioned;
-from intermediate
+from intermediate_n3
insert into table multi_partitioned partition(p) select p, key, p
insert into table multi_partitioned partition(p=1) select key, p;
select key, key2, p from multi_partitioned order by key, key2, p;
desc formatted multi_partitioned;
-from intermediate
+from intermediate_n3
insert into table multi_partitioned partition(p) select p, key, 1
insert into table multi_partitioned partition(p=1) select key, p;
@@ -52,6 +52,6 @@ desc formatted multi_partitioned;
drop table multi_partitioned;
-drop table intermediate;
+drop table intermediate_n3;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/multi_insert_union_src.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_insert_union_src.q b/ql/src/test/queries/clientpositive/multi_insert_union_src.q
index 51b4695..782c94f 100644
--- a/ql/src/test/queries/clientpositive/multi_insert_union_src.q
+++ b/ql/src/test/queries/clientpositive/multi_insert_union_src.q
@@ -1,21 +1,21 @@
set hive.mapred.mode=nonstrict;
-drop table if exists src2;
-drop table if exists src_multi1;
-drop table if exists src_multi2;
+drop table if exists src2_n4;
+drop table if exists src_multi1_n3;
+drop table if exists src_multi2_n4;
set hive.stats.dbclass=fs;
-CREATE TABLE src2 as SELECT * FROM src;
+CREATE TABLE src2_n4 as SELECT * FROM src;
-create table src_multi1 like src;
-create table src_multi2 like src;
+create table src_multi1_n3 like src;
+create table src_multi2_n4 like src;
explain
-from (select * from src1 where key < 10 union all select * from src2 where key > 100) s
-insert overwrite table src_multi1 select key, value where key < 150 order by key
-insert overwrite table src_multi2 select key, value where key > 400 order by value;
+from (select * from src1 where key < 10 union all select * from src2_n4 where key > 100) s
+insert overwrite table src_multi1_n3 select key, value where key < 150 order by key
+insert overwrite table src_multi2_n4 select key, value where key > 400 order by value;
-from (select * from src1 where key < 10 union all select * from src2 where key > 100) s
-insert overwrite table src_multi1 select key, value where key < 150 order by key
-insert overwrite table src_multi2 select key, value where key > 400 order by value;
+from (select * from src1 where key < 10 union all select * from src2_n4 where key > 100) s
+insert overwrite table src_multi1_n3 select key, value where key < 150 order by key
+insert overwrite table src_multi2_n4 select key, value where key > 400 order by value;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n3;
+select * from src_multi2_n4;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/multigroupby_singlemr.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multigroupby_singlemr.q b/ql/src/test/queries/clientpositive/multigroupby_singlemr.q
index a4e86ad..a5f6fea 100644
--- a/ql/src/test/queries/clientpositive/multigroupby_singlemr.q
+++ b/ql/src/test/queries/clientpositive/multigroupby_singlemr.q
@@ -1,33 +1,33 @@
-CREATE TABLE TBL(C1 INT, C2 INT, C3 INT, C4 INT);
+CREATE TABLE TBL_n0(C1 INT, C2 INT, C3 INT, C4 INT);
-CREATE TABLE DEST1(d1 INT, d2 INT) STORED AS TEXTFILE;
-CREATE TABLE DEST2(d1 INT, d2 INT, d3 INT) STORED AS TEXTFILE;
-CREATE TABLE DEST3(d1 INT, d2 INT, d3 INT, d4 INT) STORED AS TEXTFILE;
+CREATE TABLE DEST1_n116(d1 INT, d2 INT) STORED AS TEXTFILE;
+CREATE TABLE DEST2_n30(d1 INT, d2 INT, d3 INT) STORED AS TEXTFILE;
+CREATE TABLE DEST3_n4(d1 INT, d2 INT, d3 INT, d4 INT) STORED AS TEXTFILE;
CREATE TABLE DEST4(d1 INT, d2 INT, d3 INT, d4 INT) STORED AS TEXTFILE;
EXPLAIN
-FROM TBL
-INSERT OVERWRITE TABLE DEST1 SELECT TBL.C1, COUNT(TBL.C2) GROUP BY TBL.C1
-INSERT OVERWRITE TABLE DEST2 SELECT TBL.C1, TBL.C2, COUNT(TBL.C3) GROUP BY TBL.C1, TBL.C2;
+FROM TBL_n0
+INSERT OVERWRITE TABLE DEST1_n116 SELECT TBL_n0.C1, COUNT(TBL_n0.C2) GROUP BY TBL_n0.C1
+INSERT OVERWRITE TABLE DEST2_n30 SELECT TBL_n0.C1, TBL_n0.C2, COUNT(TBL_n0.C3) GROUP BY TBL_n0.C1, TBL_n0.C2;
EXPLAIN
-FROM TBL
-INSERT OVERWRITE TABLE DEST1 SELECT TBL.C1, COUNT(TBL.C2) GROUP BY TBL.C1
-INSERT OVERWRITE TABLE DEST2 SELECT TBL.C1, TBL.C2, COUNT(TBL.C3) GROUP BY TBL.C2, TBL.C1;
+FROM TBL_n0
+INSERT OVERWRITE TABLE DEST1_n116 SELECT TBL_n0.C1, COUNT(TBL_n0.C2) GROUP BY TBL_n0.C1
+INSERT OVERWRITE TABLE DEST2_n30 SELECT TBL_n0.C1, TBL_n0.C2, COUNT(TBL_n0.C3) GROUP BY TBL_n0.C2, TBL_n0.C1;
EXPLAIN
-FROM TBL
-INSERT OVERWRITE TABLE DEST3 SELECT TBL.C1, TBL.C2, TBL.C3, COUNT(TBL.C4) GROUP BY TBL.C1, TBL.C2, TBL.C3
-INSERT OVERWRITE TABLE DEST2 SELECT TBL.C1, TBL.C2, COUNT(TBL.C3) GROUP BY TBL.C1, TBL.C2;
+FROM TBL_n0
+INSERT OVERWRITE TABLE DEST3_n4 SELECT TBL_n0.C1, TBL_n0.C2, TBL_n0.C3, COUNT(TBL_n0.C4) GROUP BY TBL_n0.C1, TBL_n0.C2, TBL_n0.C3
+INSERT OVERWRITE TABLE DEST2_n30 SELECT TBL_n0.C1, TBL_n0.C2, COUNT(TBL_n0.C3) GROUP BY TBL_n0.C1, TBL_n0.C2;
EXPLAIN
-FROM TBL
-INSERT OVERWRITE TABLE DEST3 SELECT TBL.C1, TBL.C2, TBL.C3, COUNT(TBL.C4) GROUP BY TBL.C1, TBL.C2, TBL.C3
-INSERT OVERWRITE TABLE DEST4 SELECT TBL.C1, TBL.C2, TBL.C3, COUNT(TBL.C4) GROUP BY TBL.C1, TBL.C3, TBL.C2;
+FROM TBL_n0
+INSERT OVERWRITE TABLE DEST3_n4 SELECT TBL_n0.C1, TBL_n0.C2, TBL_n0.C3, COUNT(TBL_n0.C4) GROUP BY TBL_n0.C1, TBL_n0.C2, TBL_n0.C3
+INSERT OVERWRITE TABLE DEST4 SELECT TBL_n0.C1, TBL_n0.C2, TBL_n0.C3, COUNT(TBL_n0.C4) GROUP BY TBL_n0.C1, TBL_n0.C3, TBL_n0.C2;
EXPLAIN
-FROM TBL
-INSERT OVERWRITE TABLE DEST3 SELECT TBL.C1, TBL.C2, TBL.C3, COUNT(TBL.C4) GROUP BY TBL.C1, TBL.C2, TBL.C3
-INSERT OVERWRITE TABLE DEST2 SELECT TBL.C1, TBL.C2, COUNT(TBL.C3) GROUP BY TBL.C1, TBL.C2
-INSERT OVERWRITE TABLE DEST1 SELECT TBL.C1, COUNT(TBL.C2) GROUP BY TBL.C1;
+FROM TBL_n0
+INSERT OVERWRITE TABLE DEST3_n4 SELECT TBL_n0.C1, TBL_n0.C2, TBL_n0.C3, COUNT(TBL_n0.C4) GROUP BY TBL_n0.C1, TBL_n0.C2, TBL_n0.C3
+INSERT OVERWRITE TABLE DEST2_n30 SELECT TBL_n0.C1, TBL_n0.C2, COUNT(TBL_n0.C3) GROUP BY TBL_n0.C1, TBL_n0.C2
+INSERT OVERWRITE TABLE DEST1_n116 SELECT TBL_n0.C1, COUNT(TBL_n0.C2) GROUP BY TBL_n0.C1;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/named_column_join.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/named_column_join.q b/ql/src/test/queries/clientpositive/named_column_join.q
index 6de9325..7a0cc58 100644
--- a/ql/src/test/queries/clientpositive/named_column_join.q
+++ b/ql/src/test/queries/clientpositive/named_column_join.q
@@ -1,52 +1,52 @@
set hive.mapred.mode=nonstrict;
-- SORT_QUERY_RESULTS
-create table t (a int, b int);
+create table t_n8 (a int, b int);
-insert into t values (1,2),(2,1),(3,4),(4,3),(3,3),(null,null),(null,1),(2,null);
+insert into t_n8 values (1,2),(2,1),(3,4),(4,3),(3,3),(null,null),(null,1),(2,null);
-explain select * from t t1 join t t2 using (a);
+explain select * from t_n8 t1 join t_n8 t2 using (a);
-select * from t t1 join t t2 using (a);
+select * from t_n8 t1 join t_n8 t2 using (a);
-select * from t t1 join t t2 using (a,b);
+select * from t_n8 t1 join t_n8 t2 using (a,b);
-select t1.a,t2.b,t1.b,t2.a from t t1 join t t2 using (a);
+select t1.a,t2.b,t1.b,t2.a from t_n8 t1 join t_n8 t2 using (a);
-select * from t t1 left outer join t t2 using (a,b);
+select * from t_n8 t1 left outer join t_n8 t2 using (a,b);
-select t1.a,t1.b from t t1 right outer join t t2 on (t1.a=t2.a and t1.b=t2.b);
+select t1.a,t1.b from t_n8 t1 right outer join t_n8 t2 on (t1.a=t2.a and t1.b=t2.b);
-select * from t t1 right outer join t t2 using (a,b);
+select * from t_n8 t1 right outer join t_n8 t2 using (a,b);
-select * from t t1 inner join t t2 using (a,b);
+select * from t_n8 t1 inner join t_n8 t2 using (a,b);
-select * from t t1 left outer join t t2 using (b);
+select * from t_n8 t1 left outer join t_n8 t2 using (b);
-select * from t t1 right outer join t t2 using (b);
+select * from t_n8 t1 right outer join t_n8 t2 using (b);
-select * from t t1 inner join t t2 using (b);
+select * from t_n8 t1 inner join t_n8 t2 using (b);
-drop view v;
+drop view v_n4;
-create view v as select * from t t1 join t t2 using (a,b);
+create view v_n4 as select * from t_n8 t1 join t_n8 t2 using (a,b);
-desc formatted v;
+desc formatted v_n4;
-select * from v;
+select * from v_n4;
-drop view v;
+drop view v_n4;
-create view v as select * from t t1 right outer join t t2 using (b,a);
+create view v_n4 as select * from t_n8 t1 right outer join t_n8 t2 using (b,a);
-desc formatted v;
+desc formatted v_n4;
-select * from v;
+select * from v_n4;
-select * from (select t1.b b from t t1 inner join t t2 using (b)) t3 join t t4 using(b);
+select * from (select t1.b b from t_n8 t1 inner join t_n8 t2 using (b)) t3 join t_n8 t4 using(b);
-select * from (select t2.a a from t t1 inner join t t2 using (b)) t3 join t t4 using(a);
+select * from (select t2.a a from t_n8 t1 inner join t_n8 t2 using (b)) t3 join t_n8 t4 using(a);
-create table tt as select * from (select t2.a a from t t1 inner join t t2 using (b)) t3 join t t4 using(a);
+create table tt_n0 as select * from (select t2.a a from t_n8 t1 inner join t_n8 t2 using (b)) t3 join t_n8 t4 using(a);
-desc formatted tt;
+desc formatted tt_n0;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/nested_column_pruning.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/nested_column_pruning.q b/ql/src/test/queries/clientpositive/nested_column_pruning.q
index ab5e16b..5b41f18 100644
--- a/ql/src/test/queries/clientpositive/nested_column_pruning.q
+++ b/ql/src/test/queries/clientpositive/nested_column_pruning.q
@@ -7,12 +7,12 @@ set hive.exec.dynamic.partition.mode = nonstrict;
set hive.strict.checks.cartesian.product=false;
-- First, create source tables
-DROP TABLE IF EXISTS dummy;
-CREATE TABLE dummy (i int);
-INSERT INTO TABLE dummy VALUES (42);
+DROP TABLE IF EXISTS dummy_n5;
+CREATE TABLE dummy_n5 (i int);
+INSERT INTO TABLE dummy_n5 VALUES (42);
-DROP TABLE IF EXISTS nested_tbl_1;
-CREATE TABLE nested_tbl_1 (
+DROP TABLE IF EXISTS nested_tbl_1_n1;
+CREATE TABLE nested_tbl_1_n1 (
a int,
s1 struct<f1: boolean, f2: string, f3: struct<f4: int, f5: double>, f6: int>,
s2 struct<f7: string, f8: struct<f9 : boolean, f10: array<int>, f11: map<string, boolean>>>,
@@ -22,7 +22,7 @@ CREATE TABLE nested_tbl_1 (
s6 map<string, struct<f20:array<struct<f21:struct<f22:int>>>>>
) STORED AS PARQUET;
-INSERT INTO TABLE nested_tbl_1 SELECT
+INSERT INTO TABLE nested_tbl_1_n1 SELECT
1, named_struct('f1', false, 'f2', 'foo', 'f3', named_struct('f4', 4, 'f5', cast(5.0 as double)), 'f6', 4),
named_struct('f7', 'f7', 'f8', named_struct('f9', true, 'f10', array(10, 11), 'f11', map('key1', true, 'key2', false))),
named_struct('f12', array(named_struct('f13', 'foo', 'f14', 14), named_struct('f13', 'bar', 'f14', 28))),
@@ -30,12 +30,12 @@ INSERT INTO TABLE nested_tbl_1 SELECT
named_struct('f16', array(named_struct('f17', 'foo', 'f18', named_struct('f19', 14)), named_struct('f17', 'bar', 'f18', named_struct('f19', 28)))),
map('key1', named_struct('f20', array(named_struct('f21', named_struct('f22', 1)))),
'key2', named_struct('f20', array(named_struct('f21', named_struct('f22', 2)))))
-FROM dummy;
+FROM dummy_n5;
-DROP TABLE IF EXISTS nested_tbl_2;
-CREATE TABLE nested_tbl_2 LIKE nested_tbl_1;
+DROP TABLE IF EXISTS nested_tbl_2_n1;
+CREATE TABLE nested_tbl_2_n1 LIKE nested_tbl_1_n1;
-INSERT INTO TABLE nested_tbl_2 SELECT
+INSERT INTO TABLE nested_tbl_2_n1 SELECT
2, named_struct('f1', true, 'f2', 'bar', 'f3', named_struct('f4', 4, 'f5', cast(6.5 as double)), 'f6', 4),
named_struct('f7', 'f72', 'f8', named_struct('f9', false, 'f10', array(20, 22), 'f11', map('key3', true, 'key4', false))),
named_struct('f12', array(named_struct('f13', 'bar', 'f14', 28), named_struct('f13', 'foo', 'f14', 56))),
@@ -43,175 +43,175 @@ INSERT INTO TABLE nested_tbl_2 SELECT
named_struct('f16', array(named_struct('f17', 'bar', 'f18', named_struct('f19', 28)), named_struct('f17', 'foo', 'f18', named_struct('f19', 56)))),
map('key3', named_struct('f20', array(named_struct('f21', named_struct('f22', 3)))),
'key4', named_struct('f20', array(named_struct('f21', named_struct('f22', 4)))))
-FROM dummy;
+FROM dummy_n5;
-- Testing only select statements
-EXPLAIN SELECT a FROM nested_tbl_1;
-SELECT a FROM nested_tbl_1;
+EXPLAIN SELECT a FROM nested_tbl_1_n1;
+SELECT a FROM nested_tbl_1_n1;
-EXPLAIN SELECT s1.f1 FROM nested_tbl_1;
-SELECT s1.f1 FROM nested_tbl_1;
+EXPLAIN SELECT s1.f1 FROM nested_tbl_1_n1;
+SELECT s1.f1 FROM nested_tbl_1_n1;
-EXPLAIN SELECT s1.f1, s1.f2 FROM nested_tbl_1;
-SELECT s1.f1, s1.f2 FROM nested_tbl_1;
+EXPLAIN SELECT s1.f1, s1.f2 FROM nested_tbl_1_n1;
+SELECT s1.f1, s1.f2 FROM nested_tbl_1_n1;
-- In this case 's1.f3' and 's1.f3.f4' should be merged
-EXPLAIN SELECT s1.f3, s1.f3.f4 FROM nested_tbl_1;
-SELECT s1.f3, s1.f3.f4 FROM nested_tbl_1;
+EXPLAIN SELECT s1.f3, s1.f3.f4 FROM nested_tbl_1_n1;
+SELECT s1.f3, s1.f3.f4 FROM nested_tbl_1_n1;
-- Testing select array and index shifting
-EXPLAIN SELECT s1.f3.f5 FROM nested_tbl_1;
-SELECT s1.f3.f5 FROM nested_tbl_1;
+EXPLAIN SELECT s1.f3.f5 FROM nested_tbl_1_n1;
+SELECT s1.f3.f5 FROM nested_tbl_1_n1;
-- Testing select from multiple structs
-EXPLAIN SELECT s1.f3.f4, s2.f8.f9 FROM nested_tbl_1;
-SELECT s1.f3.f4, s2.f8.f9 FROM nested_tbl_1;
+EXPLAIN SELECT s1.f3.f4, s2.f8.f9 FROM nested_tbl_1_n1;
+SELECT s1.f3.f4, s2.f8.f9 FROM nested_tbl_1_n1;
-- Testing select with filter
-EXPLAIN SELECT s1.f2 FROM nested_tbl_1 WHERE s1.f1 = FALSE;
-SELECT s1.f2 FROM nested_tbl_1 WHERE s1.f1 = FALSE;
+EXPLAIN SELECT s1.f2 FROM nested_tbl_1_n1 WHERE s1.f1 = FALSE;
+SELECT s1.f2 FROM nested_tbl_1_n1 WHERE s1.f1 = FALSE;
-EXPLAIN SELECT s1.f3.f5 FROM nested_tbl_1 WHERE s1.f3.f4 = 4;
-SELECT s1.f3.f5 FROM nested_tbl_1 WHERE s1.f3.f4 = 4;
+EXPLAIN SELECT s1.f3.f5 FROM nested_tbl_1_n1 WHERE s1.f3.f4 = 4;
+SELECT s1.f3.f5 FROM nested_tbl_1_n1 WHERE s1.f3.f4 = 4;
-EXPLAIN SELECT s2.f8 FROM nested_tbl_1 WHERE s1.f2 = 'foo' AND size(s2.f8.f10) > 1 AND s2.f8.f11['key1'] = TRUE;
-SELECT s2.f8 FROM nested_tbl_1 WHERE s1.f2 = 'foo' AND size(s2.f8.f10) > 1 AND s2.f8.f11['key1'] = TRUE;
+EXPLAIN SELECT s2.f8 FROM nested_tbl_1_n1 WHERE s1.f2 = 'foo' AND size(s2.f8.f10) > 1 AND s2.f8.f11['key1'] = TRUE;
+SELECT s2.f8 FROM nested_tbl_1_n1 WHERE s1.f2 = 'foo' AND size(s2.f8.f10) > 1 AND s2.f8.f11['key1'] = TRUE;
-- Testing lateral view
-EXPLAIN SELECT col1, col2 FROM nested_tbl_1
+EXPLAIN SELECT col1, col2 FROM nested_tbl_1_n1
LATERAL VIEW explode(s2.f8.f10) tbl1 AS col1
LATERAL VIEW explode(s3.f12) tbl2 AS col2;
-SELECT col1, col2 FROM nested_tbl_1
+SELECT col1, col2 FROM nested_tbl_1_n1
LATERAL VIEW explode(s2.f8.f10) tbl1 AS col1
LATERAL VIEW explode(s3.f12) tbl2 AS col2;
-- Testing UDFs
-EXPLAIN SELECT pmod(s2.f8.f10[1], s1.f3.f4) FROM nested_tbl_1;
-SELECT pmod(s2.f8.f10[1], s1.f3.f4) FROM nested_tbl_1;
+EXPLAIN SELECT pmod(s2.f8.f10[1], s1.f3.f4) FROM nested_tbl_1_n1;
+SELECT pmod(s2.f8.f10[1], s1.f3.f4) FROM nested_tbl_1_n1;
-- Testing aggregations
-EXPLAIN SELECT s1.f3.f5, count(s1.f3.f4) FROM nested_tbl_1 GROUP BY s1.f3.f5;
-SELECT s1.f3.f5, count(s1.f3.f4) FROM nested_tbl_1 GROUP BY s1.f3.f5;
+EXPLAIN SELECT s1.f3.f5, count(s1.f3.f4) FROM nested_tbl_1_n1 GROUP BY s1.f3.f5;
+SELECT s1.f3.f5, count(s1.f3.f4) FROM nested_tbl_1_n1 GROUP BY s1.f3.f5;
-EXPLAIN SELECT s1.f3, count(s1.f3.f4) FROM nested_tbl_1 GROUP BY s1.f3;
-SELECT s1.f3, count(s1.f3.f4) FROM nested_tbl_1 GROUP BY s1.f3;
+EXPLAIN SELECT s1.f3, count(s1.f3.f4) FROM nested_tbl_1_n1 GROUP BY s1.f3;
+SELECT s1.f3, count(s1.f3.f4) FROM nested_tbl_1_n1 GROUP BY s1.f3;
-EXPLAIN SELECT s1.f3, count(s1.f3.f4) FROM nested_tbl_1 GROUP BY s1.f3 ORDER BY s1.f3;
-SELECT s1.f3, count(s1.f3.f4) FROM nested_tbl_1 GROUP BY s1.f3 ORDER BY s1.f3;
+EXPLAIN SELECT s1.f3, count(s1.f3.f4) FROM nested_tbl_1_n1 GROUP BY s1.f3 ORDER BY s1.f3;
+SELECT s1.f3, count(s1.f3.f4) FROM nested_tbl_1_n1 GROUP BY s1.f3 ORDER BY s1.f3;
-- Testing joins
EXPLAIN SELECT t1.s1.f3.f5, t2.s2.f8
-FROM nested_tbl_1 t1 JOIN nested_tbl_2 t2
+FROM nested_tbl_1_n1 t1 JOIN nested_tbl_2_n1 t2
ON t1.s1.f3.f4 = t2.s1.f6
WHERE t2.s2.f8.f9 == FALSE;
SELECT t1.s1.f3.f5, t2.s2.f8
-FROM nested_tbl_1 t1 JOIN nested_tbl_2 t2
+FROM nested_tbl_1_n1 t1 JOIN nested_tbl_2_n1 t2
ON t1.s1.f3.f4 = t2.s1.f6
WHERE t2.s2.f8.f9 == FALSE;
EXPLAIN SELECT t1.s1.f3.f5, t2.s2.f8
-FROM nested_tbl_1 t1 JOIN nested_tbl_1 t2
+FROM nested_tbl_1_n1 t1 JOIN nested_tbl_1_n1 t2
ON t1.s1.f3.f4 = t2.s1.f6
WHERE t2.s2.f8.f9 == TRUE;
SELECT t1.s1.f3.f5, t2.s2.f8
-FROM nested_tbl_1 t1 JOIN nested_tbl_1 t2
+FROM nested_tbl_1_n1 t1 JOIN nested_tbl_1_n1 t2
ON t1.s1.f3.f4 = t2.s1.f6
WHERE t2.s2.f8.f9 == TRUE;
EXPLAIN SELECT t1.s1.f3.f5
-FROM nested_tbl_1 t1 LEFT SEMI JOIN nested_tbl_1 t2
+FROM nested_tbl_1_n1 t1 LEFT SEMI JOIN nested_tbl_1_n1 t2
ON t1.s1.f3.f4 = t2.s1.f6 AND t2.s2.f8.f9 == TRUE;
SELECT t1.s1.f3.f5
-FROM nested_tbl_1 t1 LEFT SEMI JOIN nested_tbl_1 t2
+FROM nested_tbl_1_n1 t1 LEFT SEMI JOIN nested_tbl_1_n1 t2
ON t1.s1.f3.f4 = t2.s1.f6 AND t2.s2.f8.f9 == TRUE;
EXPLAIN SELECT t1.s1.f3.f5
-FROM nested_tbl_1 t1 LEFT SEMI JOIN nested_tbl_1 t2
+FROM nested_tbl_1_n1 t1 LEFT SEMI JOIN nested_tbl_1_n1 t2
ON t1.s1.f1 <> t2.s2.f8.f9;
SELECT t1.s1.f3.f5
-FROM nested_tbl_1 t1 LEFT SEMI JOIN nested_tbl_1 t2
+FROM nested_tbl_1_n1 t1 LEFT SEMI JOIN nested_tbl_1_n1 t2
ON t1.s1.f1 <> t2.s2.f8.f9;
EXPLAIN SELECT t1.s1.f3.f5
-FROM nested_tbl_1 t1 LEFT SEMI JOIN nested_tbl_1 t2
+FROM nested_tbl_1_n1 t1 LEFT SEMI JOIN nested_tbl_1_n1 t2
ON t1.s1.f3.f4 = t2.s1.f6 AND t1.s1.f1 <> t2.s2.f8.f9;
SELECT t1.s1.f3.f5
-FROM nested_tbl_1 t1 LEFT SEMI JOIN nested_tbl_1 t2
+FROM nested_tbl_1_n1 t1 LEFT SEMI JOIN nested_tbl_1_n1 t2
ON t1.s1.f3.f4 = t2.s1.f6 AND t1.s1.f1 <> t2.s2.f8.f9;
-- Testing insert with aliases
-DROP TABLE IF EXISTS nested_tbl_3;
-CREATE TABLE nested_tbl_3 (f1 boolean, f2 string) PARTITIONED BY (f3 int) STORED AS PARQUET;
+DROP TABLE IF EXISTS nested_tbl_3_n1;
+CREATE TABLE nested_tbl_3_n1 (f1 boolean, f2 string) PARTITIONED BY (f3 int) STORED AS PARQUET;
-INSERT OVERWRITE TABLE nested_tbl_3 PARTITION(f3)
+INSERT OVERWRITE TABLE nested_tbl_3_n1 PARTITION(f3)
SELECT s1.f1 AS f1, S1.f2 AS f2, s1.f6 AS f3
-FROM nested_tbl_1;
+FROM nested_tbl_1_n1;
-SELECT * FROM nested_tbl_3;
+SELECT * FROM nested_tbl_3_n1;
-- Testing select struct field from elements in array or map
EXPLAIN
SELECT count(s1.f6), s3.f12[0].f14
-FROM nested_tbl_1
+FROM nested_tbl_1_n1
GROUP BY s3.f12[0].f14;
SELECT count(s1.f6), s3.f12[0].f14
-FROM nested_tbl_1
+FROM nested_tbl_1_n1
GROUP BY s3.f12[0].f14;
EXPLAIN
SELECT count(s1.f6), s4['key1'].f15
-FROM nested_tbl_1
+FROM nested_tbl_1_n1
GROUP BY s4['key1'].f15;
SELECT count(s1.f6), s4['key1'].f15
-FROM nested_tbl_1
+FROM nested_tbl_1_n1
GROUP BY s4['key1'].f15;
EXPLAIN
SELECT count(s1.f6), s5.f16[0].f18.f19
-FROM nested_tbl_1
+FROM nested_tbl_1_n1
GROUP BY s5.f16[0].f18.f19;
SELECT count(s1.f6), s5.f16[0].f18.f19
-FROM nested_tbl_1
+FROM nested_tbl_1_n1
GROUP BY s5.f16[0].f18.f19;
EXPLAIN
SELECT count(s1.f6), s5.f16.f18.f19
-FROM nested_tbl_1
+FROM nested_tbl_1_n1
GROUP BY s5.f16.f18.f19;
SELECT count(s1.f6), s5.f16.f18.f19
-FROM nested_tbl_1
+FROM nested_tbl_1_n1
GROUP BY s5.f16.f18.f19;
EXPLAIN
SELECT count(s1.f6), s6['key1'].f20[0].f21.f22
-FROM nested_tbl_1
+FROM nested_tbl_1_n1
GROUP BY s6['key1'].f20[0].f21.f22;
SELECT count(s1.f6), s6['key1'].f20[0].f21.f22
-FROM nested_tbl_1
+FROM nested_tbl_1_n1
GROUP BY s6['key1'].f20[0].f21.f22;
EXPLAIN
SELECT count(s1.f6), s6['key1'].f20.f21.f22
-FROM nested_tbl_1
+FROM nested_tbl_1_n1
GROUP BY s6['key1'].f20.f21.f22;
SELECT count(s1.f6), s6['key1'].f20.f21.f22
-FROM nested_tbl_1
+FROM nested_tbl_1_n1
GROUP BY s6['key1'].f20.f21.f22;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/newline.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/newline.q b/ql/src/test/queries/clientpositive/newline.q
index e7dcb62..afe21ba 100644
--- a/ql/src/test/queries/clientpositive/newline.q
+++ b/ql/src/test/queries/clientpositive/newline.q
@@ -4,56 +4,56 @@ set hive.transform.escape.input=true;
-- SORT_QUERY_RESULTS
-create table tmp_tmp(key string, value string) stored as rcfile;
-insert overwrite table tmp_tmp
+create table tmp_tmp_n0(key string, value string) stored as rcfile;
+insert overwrite table tmp_tmp_n0
SELECT TRANSFORM(key, value) USING
'python newline.py' AS key, value FROM src limit 6;
-select * from tmp_tmp;
+select * from tmp_tmp_n0;
-drop table tmp_tmp;
+drop table tmp_tmp_n0;
add file ../../data/scripts/escapednewline.py;
add file ../../data/scripts/escapedtab.py;
add file ../../data/scripts/doubleescapedtab.py;
add file ../../data/scripts/escapedcarriagereturn.py;
-create table tmp_tmp(key string, value string) stored as rcfile;
-insert overwrite table tmp_tmp
+create table tmp_tmp_n0(key string, value string) stored as rcfile;
+insert overwrite table tmp_tmp_n0
SELECT TRANSFORM(key, value) USING
'python escapednewline.py' AS key, value FROM src limit 5;
-select * from tmp_tmp;
+select * from tmp_tmp_n0;
SELECT TRANSFORM(key, value) USING
-'cat' AS (key, value) FROM tmp_tmp;
+'cat' AS (key, value) FROM tmp_tmp_n0;
-insert overwrite table tmp_tmp
+insert overwrite table tmp_tmp_n0
SELECT TRANSFORM(key, value) USING
'python escapedcarriagereturn.py' AS key, value FROM src limit 5;
-select * from tmp_tmp;
+select * from tmp_tmp_n0;
SELECT TRANSFORM(key, value) USING
-'cat' AS (key, value) FROM tmp_tmp;
+'cat' AS (key, value) FROM tmp_tmp_n0;
-insert overwrite table tmp_tmp
+insert overwrite table tmp_tmp_n0
SELECT TRANSFORM(key, value) USING
'python escapedtab.py' AS key, value FROM src limit 5;
-select * from tmp_tmp;
+select * from tmp_tmp_n0;
SELECT TRANSFORM(key, value) USING
-'cat' AS (key, value) FROM tmp_tmp;
+'cat' AS (key, value) FROM tmp_tmp_n0;
-insert overwrite table tmp_tmp
+insert overwrite table tmp_tmp_n0
SELECT TRANSFORM(key, value) USING
'python doubleescapedtab.py' AS key, value FROM src limit 5;
-select * from tmp_tmp;
+select * from tmp_tmp_n0;
SELECT TRANSFORM(key, value) USING
-'cat' AS (key, value) FROM tmp_tmp;
+'cat' AS (key, value) FROM tmp_tmp_n0;
SELECT key FROM (SELECT TRANSFORM ('a\tb', 'c') USING 'cat' AS (key, value) FROM src limit 1)a ORDER BY key ASC;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/notable_alias1.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/notable_alias1.q b/ql/src/test/queries/clientpositive/notable_alias1.q
index 90230d5..3b720bf 100644
--- a/ql/src/test/queries/clientpositive/notable_alias1.q
+++ b/ql/src/test/queries/clientpositive/notable_alias1.q
@@ -1,11 +1,11 @@
set hive.mapred.mode=nonstrict;
-CREATE TABLE dest1(dummy STRING, key INT, value DOUBLE) STORED AS TEXTFILE;
+CREATE TABLE dest1_n4(dummy STRING, key INT, value DOUBLE) STORED AS TEXTFILE;
EXPLAIN
FROM src
-INSERT OVERWRITE TABLE dest1 SELECT '1234', key, count(1) WHERE src.key < 100 group by key;
+INSERT OVERWRITE TABLE dest1_n4 SELECT '1234', key, count(1) WHERE src.key < 100 group by key;
FROM src
-INSERT OVERWRITE TABLE dest1 SELECT '1234', key, count(1) WHERE src.key < 100 group by key;
+INSERT OVERWRITE TABLE dest1_n4 SELECT '1234', key, count(1) WHERE src.key < 100 group by key;
-SELECT dest1.* FROM dest1;
+SELECT dest1_n4.* FROM dest1_n4;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/notable_alias2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/notable_alias2.q b/ql/src/test/queries/clientpositive/notable_alias2.q
index 2f32b32..c6814dc 100644
--- a/ql/src/test/queries/clientpositive/notable_alias2.q
+++ b/ql/src/test/queries/clientpositive/notable_alias2.q
@@ -1,11 +1,11 @@
set hive.mapred.mode=nonstrict;
-CREATE TABLE dest1(dummy STRING, key INT, value DOUBLE) STORED AS TEXTFILE;
+CREATE TABLE dest1_n50(dummy STRING, key INT, value DOUBLE) STORED AS TEXTFILE;
EXPLAIN
FROM src
-INSERT OVERWRITE TABLE dest1 SELECT '1234', src.key, count(1) WHERE key < 100 group by src.key;
+INSERT OVERWRITE TABLE dest1_n50 SELECT '1234', src.key, count(1) WHERE key < 100 group by src.key;
FROM src
-INSERT OVERWRITE TABLE dest1 SELECT '1234', src.key, count(1) WHERE key < 100 group by src.key;
+INSERT OVERWRITE TABLE dest1_n50 SELECT '1234', src.key, count(1) WHERE key < 100 group by src.key;
-SELECT dest1.* FROM dest1;
+SELECT dest1_n50.* FROM dest1_n50;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/notable_alias3.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/notable_alias3.q b/ql/src/test/queries/clientpositive/notable_alias3.q
index aa79674..414ba24 100644
--- a/ql/src/test/queries/clientpositive/notable_alias3.q
+++ b/ql/src/test/queries/clientpositive/notable_alias3.q
@@ -1,4 +1,5 @@
-CREATE TABLE dest1(c string, key INT, value DOUBLE) STORED AS TEXTFILE;
+--! qt:dataset:src
+CREATE TABLE dest1_n102(c string, key INT, value DOUBLE) STORED AS TEXTFILE;
FROM src
-INSERT OVERWRITE TABLE dest1 SELECT '1234', src.key, sum(src.value) WHERE src.key < 100 group by key;
\ No newline at end of file
+INSERT OVERWRITE TABLE dest1_n102 SELECT '1234', src.key, sum(src.value) WHERE src.key < 100 group by key;
\ No newline at end of file
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/nullMap.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/nullMap.q b/ql/src/test/queries/clientpositive/nullMap.q
index f272bb9..28d9e2d 100644
--- a/ql/src/test/queries/clientpositive/nullMap.q
+++ b/ql/src/test/queries/clientpositive/nullMap.q
@@ -1,6 +1,6 @@
SET hive.vectorized.execution.enabled=false;
-create table map_txt (
+create table map_txt_n0 (
id int,
content map<int,string>
)
@@ -9,8 +9,8 @@ null defined as '\\N'
stored as textfile
;
-LOAD DATA LOCAL INPATH '../../data/files/mapNull.txt' INTO TABLE map_txt;
+LOAD DATA LOCAL INPATH '../../data/files/mapNull.txt' INTO TABLE map_txt_n0;
-select * from map_txt;
+select * from map_txt_n0;
-select id, map_keys(content) from map_txt;
+select id, map_keys(content) from map_txt_n0;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/null_column.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/null_column.q b/ql/src/test/queries/clientpositive/null_column.q
index 4b43d60..bc2c544 100644
--- a/ql/src/test/queries/clientpositive/null_column.q
+++ b/ql/src/test/queries/clientpositive/null_column.q
@@ -7,9 +7,9 @@ load data local inpath '../../data/files/test.dat' overwrite into table temp_nul
select null, null from temp_null;
-create table tt(a int, b string);
-insert overwrite table tt select null, null from temp_null;
-select * from tt;
+create table tt_n1(a int, b string);
+insert overwrite table tt_n1 select null, null from temp_null;
+select * from tt_n1;
create table tt_b(a int, b string) row format serde "org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe";
insert overwrite table tt_b select null, null from temp_null;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/nullability_transitive_inference.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/nullability_transitive_inference.q b/ql/src/test/queries/clientpositive/nullability_transitive_inference.q
index ff0ce3a..ebde989 100644
--- a/ql/src/test/queries/clientpositive/nullability_transitive_inference.q
+++ b/ql/src/test/queries/clientpositive/nullability_transitive_inference.q
@@ -6,35 +6,35 @@ set hive.strict.checks.cartesian.product=false;
set hive.stats.fetch.column.stats=true;
set hive.materializedview.rewriting=true;
-create table emps (
+create table emps_n6 (
empid int,
deptno int,
name varchar(256),
salary float,
commission int)
stored as orc TBLPROPERTIES ('transactional'='true');
-insert into emps values (100, 10, 'Bill', 10000, 1000), (200, 20, 'Eric', 8000, 500),
+insert into emps_n6 values (100, 10, 'Bill', 10000, 1000), (200, 20, 'Eric', 8000, 500),
(150, 10, 'Sebastian', 7000, null), (110, 10, 'Theodore', 10000, 250), (110, 10, 'Bill', 10000, 250);
-analyze table emps compute statistics for columns;
+analyze table emps_n6 compute statistics for columns;
-create table depts (
+create table depts_n5 (
deptno int,
name varchar(256),
locationid int)
stored as orc TBLPROPERTIES ('transactional'='true');
-insert into depts values (10, 'Sales', 10), (30, 'Marketing', null), (20, 'HR', 20);
-analyze table depts compute statistics for columns;
+insert into depts_n5 values (10, 'Sales', 10), (30, 'Marketing', null), (20, 'HR', 20);
+analyze table depts_n5 compute statistics for columns;
-alter table emps add constraint pk1 primary key (empid) disable novalidate rely;
-alter table depts add constraint pk2 primary key (deptno) disable novalidate rely;
+alter table emps_n6 add constraint pk1 primary key (empid) disable novalidate rely;
+alter table depts_n5 add constraint pk2 primary key (deptno) disable novalidate rely;
-alter table emps add constraint fk1 foreign key (deptno) references depts(deptno) disable novalidate rely;
+alter table emps_n6 add constraint fk1 foreign key (deptno) references depts_n5(deptno) disable novalidate rely;
explain
-select empid from emps
-join depts using (deptno) where depts.deptno >= 20
-group by empid, depts.deptno;
+select empid from emps_n6
+join depts_n5 using (deptno) where depts_n5.deptno >= 20
+group by empid, depts_n5.deptno;
-select empid from emps
-join depts using (deptno) where depts.deptno >= 20
-group by empid, depts.deptno;
+select empid from emps_n6
+join depts_n5 using (deptno) where depts_n5.deptno >= 20
+group by empid, depts_n5.deptno;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/nullformatCTAS.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/nullformatCTAS.q b/ql/src/test/queries/clientpositive/nullformatCTAS.q
index d077981..093742f 100644
--- a/ql/src/test/queries/clientpositive/nullformatCTAS.q
+++ b/ql/src/test/queries/clientpositive/nullformatCTAS.q
@@ -1,15 +1,15 @@
-- base table with null data
-DROP TABLE IF EXISTS base_tab;
-CREATE TABLE base_tab(a STRING, b STRING, c STRING, d STRING) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../../data/files/null.txt' INTO TABLE base_tab;
-DESCRIBE EXTENDED base_tab;
+DROP TABLE IF EXISTS base_tab_n2;
+CREATE TABLE base_tab_n2(a STRING, b STRING, c STRING, d STRING) STORED AS TEXTFILE;
+LOAD DATA LOCAL INPATH '../../data/files/null.txt' INTO TABLE base_tab_n2;
+DESCRIBE EXTENDED base_tab_n2;
-- table with non-default null format
DROP TABLE IF EXISTS null_tab3;
EXPLAIN CREATE TABLE null_tab3 ROW FORMAT DELIMITED NULL DEFINED AS 'fooNull'
- AS SELECT a, b FROM base_tab;
+ AS SELECT a, b FROM base_tab_n2;
CREATE TABLE null_tab3 ROW FORMAT DELIMITED NULL DEFINED AS 'fooNull'
- AS SELECT a, b FROM base_tab;
+ AS SELECT a, b FROM base_tab_n2;
DESCRIBE EXTENDED null_tab3;
SHOW CREATE TABLE null_tab3;
@@ -21,4 +21,4 @@ SELECT * FROM null_tab3;
DROP TABLE null_tab3;
-DROP TABLE base_tab;
+DROP TABLE base_tab_n2;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/nullformatdir.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/nullformatdir.q b/ql/src/test/queries/clientpositive/nullformatdir.q
index d298638..60637ab 100644
--- a/ql/src/test/queries/clientpositive/nullformatdir.q
+++ b/ql/src/test/queries/clientpositive/nullformatdir.q
@@ -1,13 +1,13 @@
-- base table with null data
-DROP TABLE IF EXISTS base_tab;
-CREATE TABLE base_tab(a STRING, b STRING, c STRING, d STRING) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../../data/files/null.txt' INTO TABLE base_tab;
-DESCRIBE EXTENDED base_tab;
+DROP TABLE IF EXISTS base_tab_n1;
+CREATE TABLE base_tab_n1(a STRING, b STRING, c STRING, d STRING) STORED AS TEXTFILE;
+LOAD DATA LOCAL INPATH '../../data/files/null.txt' INTO TABLE base_tab_n1;
+DESCRIBE EXTENDED base_tab_n1;
dfs ${system:test.dfs.mkdir} ${system:test.tmp.dir}/hive_test/nullformat/tmp;
dfs -rmr ${system:test.tmp.dir}/hive_test/nullformat/*;
INSERT OVERWRITE LOCAL DIRECTORY '${system:test.tmp.dir}/hive_test/nullformat'
- ROW FORMAT DELIMITED NULL DEFINED AS 'fooNull' SELECT a,b FROM base_tab;
+ ROW FORMAT DELIMITED NULL DEFINED AS 'fooNull' SELECT a,b FROM base_tab_n1;
dfs -cat ${system:test.tmp.dir}/hive_test/nullformat/000000_0;
-- load the exported data back into a table with same null format and verify null values
@@ -18,4 +18,4 @@ SELECT * FROM null_tab2;
dfs -rmr ${system:test.tmp.dir}/hive_test/nullformat;
-DROP TABLE base_tab;
+DROP TABLE base_tab_n1;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/nullgroup3.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/nullgroup3.q b/ql/src/test/queries/clientpositive/nullgroup3.q
index 282f6c3..af6689e 100644
--- a/ql/src/test/queries/clientpositive/nullgroup3.q
+++ b/ql/src/test/queries/clientpositive/nullgroup3.q
@@ -1,29 +1,29 @@
set hive.mapred.mode=nonstrict;
-CREATE TABLE tstparttbl(KEY STRING, VALUE STRING) PARTITIONED BY(ds string) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../../data/files/kv1.txt' INTO TABLE tstparttbl PARTITION (ds='2008-04-09');
-LOAD DATA LOCAL INPATH '../../data/files/nullfile.txt' INTO TABLE tstparttbl PARTITION (ds='2008-04-08');
+CREATE TABLE tstparttbl_n0(KEY STRING, VALUE STRING) PARTITIONED BY(ds string) STORED AS TEXTFILE;
+LOAD DATA LOCAL INPATH '../../data/files/kv1.txt' INTO TABLE tstparttbl_n0 PARTITION (ds='2008-04-09');
+LOAD DATA LOCAL INPATH '../../data/files/nullfile.txt' INTO TABLE tstparttbl_n0 PARTITION (ds='2008-04-08');
explain
-select count(1) from tstparttbl;
-select count(1) from tstparttbl;
+select count(1) from tstparttbl_n0;
+select count(1) from tstparttbl_n0;
-CREATE TABLE tstparttbl2(KEY STRING, VALUE STRING) PARTITIONED BY(ds string) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../../data/files/nullfile.txt' INTO TABLE tstparttbl2 PARTITION (ds='2008-04-09');
-LOAD DATA LOCAL INPATH '../../data/files/nullfile.txt' INTO TABLE tstparttbl2 PARTITION (ds='2008-04-08');
+CREATE TABLE tstparttbl2_n0(KEY STRING, VALUE STRING) PARTITIONED BY(ds string) STORED AS TEXTFILE;
+LOAD DATA LOCAL INPATH '../../data/files/nullfile.txt' INTO TABLE tstparttbl2_n0 PARTITION (ds='2008-04-09');
+LOAD DATA LOCAL INPATH '../../data/files/nullfile.txt' INTO TABLE tstparttbl2_n0 PARTITION (ds='2008-04-08');
explain
-select count(1) from tstparttbl2;
-select count(1) from tstparttbl2;
-DROP TABLE tstparttbl;
-CREATE TABLE tstparttbl(KEY STRING, VALUE STRING) PARTITIONED BY(ds string) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../../data/files/kv1.txt' INTO TABLE tstparttbl PARTITION (ds='2008-04-09');
-LOAD DATA LOCAL INPATH '../../data/files/nullfile.txt' INTO TABLE tstparttbl PARTITION (ds='2008-04-08');
+select count(1) from tstparttbl2_n0;
+select count(1) from tstparttbl2_n0;
+DROP TABLE tstparttbl_n0;
+CREATE TABLE tstparttbl_n0(KEY STRING, VALUE STRING) PARTITIONED BY(ds string) STORED AS TEXTFILE;
+LOAD DATA LOCAL INPATH '../../data/files/kv1.txt' INTO TABLE tstparttbl_n0 PARTITION (ds='2008-04-09');
+LOAD DATA LOCAL INPATH '../../data/files/nullfile.txt' INTO TABLE tstparttbl_n0 PARTITION (ds='2008-04-08');
explain
-select count(1) from tstparttbl;
-select count(1) from tstparttbl;
+select count(1) from tstparttbl_n0;
+select count(1) from tstparttbl_n0;
-DROP TABLE tstparttbl2;
-CREATE TABLE tstparttbl2(KEY STRING, VALUE STRING) PARTITIONED BY(ds string) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../../data/files/nullfile.txt' INTO TABLE tstparttbl2 PARTITION (ds='2008-04-09');
-LOAD DATA LOCAL INPATH '../../data/files/nullfile.txt' INTO TABLE tstparttbl2 PARTITION (ds='2008-04-08');
+DROP TABLE tstparttbl2_n0;
+CREATE TABLE tstparttbl2_n0(KEY STRING, VALUE STRING) PARTITIONED BY(ds string) STORED AS TEXTFILE;
+LOAD DATA LOCAL INPATH '../../data/files/nullfile.txt' INTO TABLE tstparttbl2_n0 PARTITION (ds='2008-04-09');
+LOAD DATA LOCAL INPATH '../../data/files/nullfile.txt' INTO TABLE tstparttbl2_n0 PARTITION (ds='2008-04-08');
explain
-select count(1) from tstparttbl2;
-select count(1) from tstparttbl2;
+select count(1) from tstparttbl2_n0;
+select count(1) from tstparttbl2_n0;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/optimize_filter_literal.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/optimize_filter_literal.q b/ql/src/test/queries/clientpositive/optimize_filter_literal.q
index 28bed09..b0ce4ae 100644
--- a/ql/src/test/queries/clientpositive/optimize_filter_literal.q
+++ b/ql/src/test/queries/clientpositive/optimize_filter_literal.q
@@ -12,38 +12,38 @@ set hive.vectorized.execution.enabled=true;
-- SORT_QUERY_RESULTS
-CREATE TABLE srcbucket_mapjoin(key int, value string) partitioned by (ds string) CLUSTERED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE;
-CREATE TABLE tab_part (key int, value string) PARTITIONED BY(ds STRING) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS ORCFILE;
-CREATE TABLE srcbucket_mapjoin_part (key int, value string) partitioned by (ds string) CLUSTERED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE;
+CREATE TABLE srcbucket_mapjoin_n21(key int, value string) partitioned by (ds string) CLUSTERED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE;
+CREATE TABLE tab_part_n13 (key int, value string) PARTITIONED BY(ds STRING) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS ORCFILE;
+CREATE TABLE srcbucket_mapjoin_part_n22 (key int, value string) partitioned by (ds string) CLUSTERED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE;
-load data local inpath '../../data/files/bmj/000000_0' INTO TABLE srcbucket_mapjoin partition(ds='2008-04-08');
-load data local inpath '../../data/files/bmj1/000001_0' INTO TABLE srcbucket_mapjoin partition(ds='2008-04-08');
+load data local inpath '../../data/files/bmj/000000_0' INTO TABLE srcbucket_mapjoin_n21 partition(ds='2008-04-08');
+load data local inpath '../../data/files/bmj1/000001_0' INTO TABLE srcbucket_mapjoin_n21 partition(ds='2008-04-08');
-load data local inpath '../../data/files/bmj/000000_0' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
-load data local inpath '../../data/files/bmj/000001_0' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
-load data local inpath '../../data/files/bmj/000002_0' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
-load data local inpath '../../data/files/bmj/000003_0' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
+load data local inpath '../../data/files/bmj/000000_0' INTO TABLE srcbucket_mapjoin_part_n22 partition(ds='2008-04-08');
+load data local inpath '../../data/files/bmj/000001_0' INTO TABLE srcbucket_mapjoin_part_n22 partition(ds='2008-04-08');
+load data local inpath '../../data/files/bmj/000002_0' INTO TABLE srcbucket_mapjoin_part_n22 partition(ds='2008-04-08');
+load data local inpath '../../data/files/bmj/000003_0' INTO TABLE srcbucket_mapjoin_part_n22 partition(ds='2008-04-08');
set hive.optimize.bucketingsorting=false;
-insert overwrite table tab_part partition (ds='2008-04-08')
-select key,value from srcbucket_mapjoin_part;
+insert overwrite table tab_part_n13 partition (ds='2008-04-08')
+select key,value from srcbucket_mapjoin_part_n22;
-analyze table tab_part partition (ds='2008-04-08') compute statistics for columns;
+analyze table tab_part_n13 partition (ds='2008-04-08') compute statistics for columns;
-CREATE TABLE tab(key int, value string) PARTITIONED BY(ds STRING) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS ORCFILE;
-insert overwrite table tab partition (ds='2008-04-08')
-select key,value from srcbucket_mapjoin;
+CREATE TABLE tab_n14(key int, value string) PARTITIONED BY(ds STRING) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS ORCFILE;
+insert overwrite table tab_n14 partition (ds='2008-04-08')
+select key,value from srcbucket_mapjoin_n21;
-analyze table tab partition (ds='2008-04-08') compute statistics for columns;
+analyze table tab_n14 partition (ds='2008-04-08') compute statistics for columns;
set hive.join.emit.interval=2;
set mapred.reduce.tasks=3;
select * from
-(select * from tab where tab.key = 0)a
+(select * from tab_n14 where tab_n14.key = 0)a
full outer join
-(select * from tab_part where tab_part.key = 98)b join tab_part c on a.key = b.key and b.key = c.key;
+(select * from tab_part_n13 where tab_part_n13.key = 98)b join tab_part_n13 c on a.key = b.key and b.key = c.key;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/optimize_join_ptp.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/optimize_join_ptp.q b/ql/src/test/queries/clientpositive/optimize_join_ptp.q
index 5807ec3..81d4f5e 100644
--- a/ql/src/test/queries/clientpositive/optimize_join_ptp.q
+++ b/ql/src/test/queries/clientpositive/optimize_join_ptp.q
@@ -1,16 +1,16 @@
set hive.mapred.mode=nonstrict;
set hive.explain.user=false;
-create table t1 (v string, k int);
-insert into t1 values ('people', 10), ('strangers', 20), ('parents', 30);
+create table t1_n97 (v string, k int);
+insert into t1_n97 values ('people', 10), ('strangers', 20), ('parents', 30);
-create table t2 (v string, k double);
-insert into t2 values ('people', 10), ('strangers', 20), ('parents', 30);
+create table t2_n60 (v string, k double);
+insert into t2_n60 values ('people', 10), ('strangers', 20), ('parents', 30);
-- should not produce exceptions
explain
-select * from t1 where t1.k in (select t2.k from t2 where t2.v='people') and t1.k<15;
+select * from t1_n97 where t1_n97.k in (select t2_n60.k from t2_n60 where t2_n60.v='people') and t1_n97.k<15;
-select * from t1 where t1.k in (select t2.k from t2 where t2.v='people') and t1.k<15;
+select * from t1_n97 where t1_n97.k in (select t2_n60.k from t2_n60 where t2_n60.v='people') and t1_n97.k<15;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_analyze.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_analyze.q b/ql/src/test/queries/clientpositive/orc_analyze.q
index 2683fdc..aea09d4 100644
--- a/ql/src/test/queries/clientpositive/orc_analyze.q
+++ b/ql/src/test/queries/clientpositive/orc_analyze.q
@@ -3,7 +3,7 @@ set hive.mapred.mode=nonstrict;
set hive.exec.submitviachild=false;
set hive.exec.submit.local.task.via.child=false;
-CREATE TABLE orc_create_people_staging (
+CREATE TABLE orc_create_people_staging_n0 (
id int,
first_name string,
last_name string,
@@ -12,14 +12,14 @@ CREATE TABLE orc_create_people_staging (
start_date timestamp,
state string);
-LOAD DATA LOCAL INPATH '../../data/files/orc_create_people.txt' OVERWRITE INTO TABLE orc_create_people_staging;
+LOAD DATA LOCAL INPATH '../../data/files/orc_create_people.txt' OVERWRITE INTO TABLE orc_create_people_staging_n0;
set hive.exec.dynamic.partition.mode=nonstrict;
set hive.stats.autogather=false;
-- non-partitioned table
-- partial scan gather
-CREATE TABLE orc_create_people (
+CREATE TABLE orc_create_people_n0 (
id int,
first_name string,
last_name string,
@@ -29,19 +29,19 @@ CREATE TABLE orc_create_people (
state string)
STORED AS orc;
-INSERT OVERWRITE TABLE orc_create_people SELECT * FROM orc_create_people_staging ORDER BY id;
+INSERT OVERWRITE TABLE orc_create_people_n0 SELECT * FROM orc_create_people_staging_n0 ORDER BY id;
set hive.stats.autogather = true;
-analyze table orc_create_people compute statistics;
-desc formatted orc_create_people;
+analyze table orc_create_people_n0 compute statistics;
+desc formatted orc_create_people_n0;
-analyze table orc_create_people compute statistics noscan;
-desc formatted orc_create_people;
+analyze table orc_create_people_n0 compute statistics noscan;
+desc formatted orc_create_people_n0;
-drop table orc_create_people;
+drop table orc_create_people_n0;
-- auto stats gather
-CREATE TABLE orc_create_people (
+CREATE TABLE orc_create_people_n0 (
id int,
first_name string,
last_name string,
@@ -51,16 +51,16 @@ CREATE TABLE orc_create_people (
state string)
STORED AS orc;
-INSERT OVERWRITE TABLE orc_create_people SELECT * FROM orc_create_people_staging ORDER BY id;
+INSERT OVERWRITE TABLE orc_create_people_n0 SELECT * FROM orc_create_people_staging_n0 ORDER BY id;
-desc formatted orc_create_people;
+desc formatted orc_create_people_n0;
-drop table orc_create_people;
+drop table orc_create_people_n0;
set hive.stats.autogather=false;
-- partitioned table
-- partial scan gather
-CREATE TABLE orc_create_people (
+CREATE TABLE orc_create_people_n0 (
id int,
first_name string,
last_name string,
@@ -70,22 +70,22 @@ CREATE TABLE orc_create_people (
PARTITIONED BY (state string)
STORED AS orc;
-INSERT OVERWRITE TABLE orc_create_people PARTITION (state)
- SELECT * FROM orc_create_people_staging ORDER BY id;
+INSERT OVERWRITE TABLE orc_create_people_n0 PARTITION (state)
+ SELECT * FROM orc_create_people_staging_n0 ORDER BY id;
set hive.stats.autogather = true;
-analyze table orc_create_people partition(state) compute statistics;
-desc formatted orc_create_people partition(state="Ca");
-desc formatted orc_create_people partition(state="Or");
+analyze table orc_create_people_n0 partition(state) compute statistics;
+desc formatted orc_create_people_n0 partition(state="Ca");
+desc formatted orc_create_people_n0 partition(state="Or");
-analyze table orc_create_people partition(state) compute statistics noscan;
-desc formatted orc_create_people partition(state="Ca");
-desc formatted orc_create_people partition(state="Or");
+analyze table orc_create_people_n0 partition(state) compute statistics noscan;
+desc formatted orc_create_people_n0 partition(state="Ca");
+desc formatted orc_create_people_n0 partition(state="Or");
-drop table orc_create_people;
+drop table orc_create_people_n0;
-- auto stats gather
-CREATE TABLE orc_create_people (
+CREATE TABLE orc_create_people_n0 (
id int,
first_name string,
last_name string,
@@ -95,18 +95,18 @@ CREATE TABLE orc_create_people (
PARTITIONED BY (state string)
STORED AS orc;
-INSERT OVERWRITE TABLE orc_create_people PARTITION (state)
- SELECT * FROM orc_create_people_staging ORDER BY id;
+INSERT OVERWRITE TABLE orc_create_people_n0 PARTITION (state)
+ SELECT * FROM orc_create_people_staging_n0 ORDER BY id;
-desc formatted orc_create_people partition(state="Ca");
-desc formatted orc_create_people partition(state="Or");
+desc formatted orc_create_people_n0 partition(state="Ca");
+desc formatted orc_create_people_n0 partition(state="Or");
-drop table orc_create_people;
+drop table orc_create_people_n0;
set hive.stats.autogather=false;
-- partitioned and bucketed table
-- partial scan gather
-CREATE TABLE orc_create_people (
+CREATE TABLE orc_create_people_n0 (
id int,
first_name string,
last_name string,
@@ -119,22 +119,22 @@ sorted by (last_name)
into 4 buckets
STORED AS orc;
-INSERT OVERWRITE TABLE orc_create_people PARTITION (state)
- SELECT * FROM orc_create_people_staging ORDER BY id;
+INSERT OVERWRITE TABLE orc_create_people_n0 PARTITION (state)
+ SELECT * FROM orc_create_people_staging_n0 ORDER BY id;
set hive.stats.autogather = true;
-analyze table orc_create_people partition(state) compute statistics;
-desc formatted orc_create_people partition(state="Ca");
-desc formatted orc_create_people partition(state="Or");
+analyze table orc_create_people_n0 partition(state) compute statistics;
+desc formatted orc_create_people_n0 partition(state="Ca");
+desc formatted orc_create_people_n0 partition(state="Or");
-analyze table orc_create_people partition(state) compute statistics noscan;
-desc formatted orc_create_people partition(state="Ca");
-desc formatted orc_create_people partition(state="Or");
+analyze table orc_create_people_n0 partition(state) compute statistics noscan;
+desc formatted orc_create_people_n0 partition(state="Ca");
+desc formatted orc_create_people_n0 partition(state="Or");
-drop table orc_create_people;
+drop table orc_create_people_n0;
-- auto stats gather
-CREATE TABLE orc_create_people (
+CREATE TABLE orc_create_people_n0 (
id int,
first_name string,
last_name string,
@@ -147,19 +147,19 @@ sorted by (last_name)
into 4 buckets
STORED AS orc;
-INSERT OVERWRITE TABLE orc_create_people PARTITION (state)
- SELECT * FROM orc_create_people_staging ORDER BY id;
+INSERT OVERWRITE TABLE orc_create_people_n0 PARTITION (state)
+ SELECT * FROM orc_create_people_staging_n0 ORDER BY id;
-desc formatted orc_create_people partition(state="Ca");
-desc formatted orc_create_people partition(state="Or");
+desc formatted orc_create_people_n0 partition(state="Ca");
+desc formatted orc_create_people_n0 partition(state="Or");
-drop table orc_create_people;
+drop table orc_create_people_n0;
set hive.stats.autogather=false;
-- create table with partitions containing text and ORC files.
-- ORC files implements StatsProvidingRecordReader but text files does not.
-- So the partition containing text file should not have statistics.
-CREATE TABLE orc_create_people (
+CREATE TABLE orc_create_people_n0 (
id int,
first_name string,
last_name string,
@@ -169,14 +169,14 @@ CREATE TABLE orc_create_people (
PARTITIONED BY (state string)
STORED AS orc;
-INSERT OVERWRITE TABLE orc_create_people PARTITION (state)
- SELECT * FROM orc_create_people_staging ORDER BY id;
+INSERT OVERWRITE TABLE orc_create_people_n0 PARTITION (state)
+ SELECT * FROM orc_create_people_staging_n0 ORDER BY id;
set hive.stats.autogather = true;
-analyze table orc_create_people partition(state) compute statistics;
-desc formatted orc_create_people partition(state="Ca");
+analyze table orc_create_people_n0 partition(state) compute statistics;
+desc formatted orc_create_people_n0 partition(state="Ca");
-analyze table orc_create_people partition(state) compute statistics noscan;
-desc formatted orc_create_people partition(state="Ca");
+analyze table orc_create_people_n0 partition(state) compute statistics noscan;
+desc formatted orc_create_people_n0 partition(state="Ca");
-drop table orc_create_people;
+drop table orc_create_people_n0;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_create.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_create.q b/ql/src/test/queries/clientpositive/orc_create.q
index 5b75f32..8e0fb6c 100644
--- a/ql/src/test/queries/clientpositive/orc_create.q
+++ b/ql/src/test/queries/clientpositive/orc_create.q
@@ -3,13 +3,13 @@ set hive.mapred.mode=nonstrict;
-- SORT_QUERY_RESULTS
DROP TABLE orc_create;
-DROP TABLE orc_create_complex;
-DROP TABLE orc_create_staging;
+DROP TABLE orc_create_complex_n1;
+DROP TABLE orc_create_staging_n1;
DROP TABLE orc_create_people_staging;
DROP TABLE orc_create_people;
DROP TABLE if exists orc_create_cprl;
-CREATE TABLE orc_create_staging (
+CREATE TABLE orc_create_staging_n1 (
str STRING,
mp MAP<STRING,STRING>,
lst ARRAY<STRING>,
@@ -19,7 +19,7 @@ CREATE TABLE orc_create_staging (
COLLECTION ITEMS TERMINATED BY ','
MAP KEYS TERMINATED BY ':';
-DESCRIBE FORMATTED orc_create_staging;
+DESCRIBE FORMATTED orc_create_staging_n1;
CREATE TABLE orc_create (key INT, value STRING)
PARTITIONED BY (ds string)
@@ -49,26 +49,26 @@ set hive.default.fileformat=TextFile;
DESCRIBE FORMATTED orc_create;
-CREATE TABLE orc_create_complex (
+CREATE TABLE orc_create_complex_n1 (
str STRING,
mp MAP<STRING,STRING>,
lst ARRAY<STRING>,
strct STRUCT<A:STRING,B:STRING>
) STORED AS ORC;
-DESCRIBE FORMATTED orc_create_complex;
+DESCRIBE FORMATTED orc_create_complex_n1;
-LOAD DATA LOCAL INPATH '../../data/files/orc_create.txt' OVERWRITE INTO TABLE orc_create_staging;
+LOAD DATA LOCAL INPATH '../../data/files/orc_create.txt' OVERWRITE INTO TABLE orc_create_staging_n1;
-SELECT * from orc_create_staging;
+SELECT * from orc_create_staging_n1;
-INSERT OVERWRITE TABLE orc_create_complex SELECT * FROM orc_create_staging;
+INSERT OVERWRITE TABLE orc_create_complex_n1 SELECT * FROM orc_create_staging_n1;
-SELECT * from orc_create_complex;
-SELECT str from orc_create_complex;
-SELECT mp from orc_create_complex;
-SELECT lst from orc_create_complex;
-SELECT strct from orc_create_complex;
+SELECT * from orc_create_complex_n1;
+SELECT str from orc_create_complex_n1;
+SELECT mp from orc_create_complex_n1;
+SELECT lst from orc_create_complex_n1;
+SELECT strct from orc_create_complex_n1;
CREATE TABLE orc_create_people_staging (
id int,
@@ -132,8 +132,8 @@ SELECT 1 from src limit 1;
SELECT * from orc_create_cprl;
DROP TABLE orc_create;
-DROP TABLE orc_create_complex;
-DROP TABLE orc_create_staging;
+DROP TABLE orc_create_complex_n1;
+DROP TABLE orc_create_staging_n1;
DROP TABLE orc_create_people_staging;
DROP TABLE orc_create_people;
DROP TABLE orc_create_cprl;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_dictionary_threshold.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_dictionary_threshold.q b/ql/src/test/queries/clientpositive/orc_dictionary_threshold.q
index 32de82a..2c619bc 100644
--- a/ql/src/test/queries/clientpositive/orc_dictionary_threshold.q
+++ b/ql/src/test/queries/clientpositive/orc_dictionary_threshold.q
@@ -5,19 +5,19 @@ set hive.exec.orc.dictionary.key.size.threshold=-1;
-- Tests that the data can be read back correctly when a string column is stored
-- without dictionary encoding
-CREATE TABLE test_orc (key STRING)
+CREATE TABLE test_orc_n5 (key STRING)
ROW FORMAT SERDE 'org.apache.hadoop.hive.ql.io.orc.OrcSerde'
STORED AS INPUTFORMAT 'org.apache.hadoop.hive.ql.io.orc.OrcInputFormat'
OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat';
-- should be single split
-INSERT OVERWRITE TABLE test_orc SELECT key FROM src TABLESAMPLE (10 ROWS);
+INSERT OVERWRITE TABLE test_orc_n5 SELECT key FROM src TABLESAMPLE (10 ROWS);
-- Test reading the column back
-SELECT * FROM test_orc;
+SELECT * FROM test_orc_n5;
-ALTER TABLE test_orc SET SERDEPROPERTIES ('orc.stripe.size' = '1');
+ALTER TABLE test_orc_n5 SET SERDEPROPERTIES ('orc.stripe.size' = '1');
CREATE TABLE src_thousand(key STRING) STORED AS TEXTFILE;
LOAD DATA LOCAL INPATH '../../data/files/kv1kv2.cogroup.txt'
@@ -33,7 +33,7 @@ set hive.exec.orc.dictionary.key.size.threshold=0.5;
-- dictionary encoded. The final stripe will have 630 out of 1000 and be
-- direct encoded.
-INSERT OVERWRITE TABLE test_orc
+INSERT OVERWRITE TABLE test_orc_n5
SELECT key FROM (
SELECT CONCAT("a", key) AS key FROM src_thousand
UNION ALL
@@ -58,4 +58,4 @@ UNION ALL
SELECT CONCAT("k", key) AS key FROM src_thousand
) a ORDER BY key LIMIT 11000;
-SELECT SUM(HASH(key)) FROM test_orc;
+SELECT SUM(HASH(key)) FROM test_orc_n5;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_diff_part_cols.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_diff_part_cols.q b/ql/src/test/queries/clientpositive/orc_diff_part_cols.q
index 9e7f688..4b14afc 100644
--- a/ql/src/test/queries/clientpositive/orc_diff_part_cols.q
+++ b/ql/src/test/queries/clientpositive/orc_diff_part_cols.q
@@ -2,7 +2,7 @@ set hive.vectorized.execution.enabled=false;
set hive.mapred.mode=nonstrict;
-- SORT_QUERY_RESULTS
-CREATE TABLE test_orc (key STRING)
+CREATE TABLE test_orc_n0 (key STRING)
PARTITIONED BY (part STRING)
ROW FORMAT SERDE 'org.apache.hadoop.hive.ql.io.orc.OrcSerde'
STORED AS INPUTFORMAT 'org.apache.hadoop.hive.ql.io.orc.OrcInputFormat'
@@ -14,10 +14,10 @@ set hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat;
-- to another partition
-- This can produce unexpected results with CombineHiveInputFormat
-INSERT OVERWRITE TABLE test_orc PARTITION (part = '1') SELECT key FROM src tablesample (5 rows);
+INSERT OVERWRITE TABLE test_orc_n0 PARTITION (part = '1') SELECT key FROM src tablesample (5 rows);
-ALTER TABLE test_orc ADD COLUMNS (cnt INT);
+ALTER TABLE test_orc_n0 ADD COLUMNS (cnt INT);
-INSERT OVERWRITE TABLE test_orc PARTITION (part = '2') SELECT key, count(*) FROM src GROUP BY key LIMIT 5;
+INSERT OVERWRITE TABLE test_orc_n0 PARTITION (part = '2') SELECT key, count(*) FROM src GROUP BY key LIMIT 5;
-SELECT * FROM test_orc;
+SELECT * FROM test_orc_n0;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_diff_part_cols2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_diff_part_cols2.q b/ql/src/test/queries/clientpositive/orc_diff_part_cols2.q
index 5fd81a4..2501f05 100644
--- a/ql/src/test/queries/clientpositive/orc_diff_part_cols2.q
+++ b/ql/src/test/queries/clientpositive/orc_diff_part_cols2.q
@@ -5,11 +5,11 @@ set hive.vectorized.execution.enabled=false;
-- SORT_QUERY_RESULTS
-CREATE TABLE test_orc (key STRING)
+CREATE TABLE test_orc_n4 (key STRING)
STORED AS ORC;
-INSERT OVERWRITE TABLE test_orc SELECT key FROM src LIMIT 5;
+INSERT OVERWRITE TABLE test_orc_n4 SELECT key FROM src LIMIT 5;
-ALTER TABLE test_orc ADD COLUMNS (value STRING);
+ALTER TABLE test_orc_n4 ADD COLUMNS (value STRING);
-SELECT * FROM test_orc;
+SELECT * FROM test_orc_n4;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_empty_files.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_empty_files.q b/ql/src/test/queries/clientpositive/orc_empty_files.q
index 2e592fa..76b9d6e 100644
--- a/ql/src/test/queries/clientpositive/orc_empty_files.q
+++ b/ql/src/test/queries/clientpositive/orc_empty_files.q
@@ -1,5 +1,5 @@
set hive.vectorized.execution.enabled=false;
-CREATE TABLE test_orc (key STRING, cnt INT)
+CREATE TABLE test_orc_n2 (key STRING, cnt INT)
CLUSTERED BY (key) INTO 3 BUCKETS
ROW FORMAT SERDE 'org.apache.hadoop.hive.ql.io.orc.OrcSerde'
STORED AS INPUTFORMAT 'org.apache.hadoop.hive.ql.io.orc.OrcInputFormat'
@@ -14,6 +14,6 @@ set hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat;
-- containing data and a file containing data followed by an empty file.
-- This can produce unexpected results with CombineHiveInputFormat
-INSERT OVERWRITE TABLE test_orc SELECT one, COUNT(*) FROM (SELECT 1 AS one FROM src) a GROUP BY one;
+INSERT OVERWRITE TABLE test_orc_n2 SELECT one, COUNT(*) FROM (SELECT 1 AS one FROM src) a GROUP BY one;
-SELECT count(*) FROM test_orc;
+SELECT count(*) FROM test_orc_n2;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_empty_strings.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_empty_strings.q b/ql/src/test/queries/clientpositive/orc_empty_strings.q
index 75309d7..9960fba 100644
--- a/ql/src/test/queries/clientpositive/orc_empty_strings.q
+++ b/ql/src/test/queries/clientpositive/orc_empty_strings.q
@@ -2,19 +2,19 @@ set hive.vectorized.execution.enabled=false;
-- SORT_QUERY_RESULTS
-CREATE TABLE test_orc (key STRING)
+CREATE TABLE test_orc_n3 (key STRING)
ROW FORMAT SERDE 'org.apache.hadoop.hive.ql.io.orc.OrcSerde'
STORED AS INPUTFORMAT 'org.apache.hadoop.hive.ql.io.orc.OrcInputFormat'
OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat';
-INSERT OVERWRITE TABLE test_orc SELECT '' FROM src tablesample (10 rows);
+INSERT OVERWRITE TABLE test_orc_n3 SELECT '' FROM src tablesample (10 rows);
-- Test reading a column which is just empty strings
-SELECT * FROM test_orc;
+SELECT * FROM test_orc_n3;
-INSERT OVERWRITE TABLE test_orc SELECT IF (key % 3 = 0, key, '') FROM src tablesample (10 rows);
+INSERT OVERWRITE TABLE test_orc_n3 SELECT IF (key % 3 = 0, key, '') FROM src tablesample (10 rows);
-- Test reading a column which has some empty strings
-SELECT * FROM test_orc;
+SELECT * FROM test_orc_n3;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_ends_with_nulls.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_ends_with_nulls.q b/ql/src/test/queries/clientpositive/orc_ends_with_nulls.q
index 9c33048..ba6dc05 100644
--- a/ql/src/test/queries/clientpositive/orc_ends_with_nulls.q
+++ b/ql/src/test/queries/clientpositive/orc_ends_with_nulls.q
@@ -11,9 +11,9 @@ ALTER TABLE test_orc SET SERDEPROPERTIES ('orc.row.index.stride' = '1000');
-- this produces the effect that the number of non-null rows between the last and second
-- to last index stride are the same (there's only two index strides)
-CREATE TABLE src_null(a STRING) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../../data/files/nulls.txt' INTO TABLE src_null;
+CREATE TABLE src_null_n0(a STRING) STORED AS TEXTFILE;
+LOAD DATA LOCAL INPATH '../../data/files/nulls.txt' INTO TABLE src_null_n0;
-INSERT OVERWRITE TABLE test_orc SELECT a FROM src_null;
+INSERT OVERWRITE TABLE test_orc SELECT a FROM src_null_n0;
SELECT * FROM test_orc LIMIT 5;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_file_dump.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_file_dump.q b/ql/src/test/queries/clientpositive/orc_file_dump.q
index 754e121..0aec810 100644
--- a/ql/src/test/queries/clientpositive/orc_file_dump.q
+++ b/ql/src/test/queries/clientpositive/orc_file_dump.q
@@ -1,7 +1,7 @@
set hive.vectorized.execution.enabled=false;
set hive.mapred.mode=nonstrict;
-CREATE TABLE staging(t tinyint,
+CREATE TABLE staging_n4(t tinyint,
si smallint,
i int,
b bigint,
@@ -15,9 +15,9 @@ CREATE TABLE staging(t tinyint,
ROW FORMAT DELIMITED FIELDS TERMINATED BY '|'
STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../../data/files/over1k' OVERWRITE INTO TABLE staging;
+LOAD DATA LOCAL INPATH '../../data/files/over1k' OVERWRITE INTO TABLE staging_n4;
-CREATE TABLE orc_ppd(t tinyint,
+CREATE TABLE orc_ppd_n0(t tinyint,
si smallint,
i int,
b bigint,
@@ -30,17 +30,17 @@ CREATE TABLE orc_ppd(t tinyint,
bin binary)
STORED AS ORC tblproperties("orc.row.index.stride" = "1000", "orc.bloom.filter.columns"="*");
-insert overwrite table orc_ppd select * from staging;
+insert overwrite table orc_ppd_n0 select * from staging_n4;
SET hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.PostExecOrcFileDump;
-select * from orc_ppd limit 1;
+select * from orc_ppd_n0 limit 1;
-alter table orc_ppd set tblproperties("orc.bloom.filter.fpp"="0.01");
+alter table orc_ppd_n0 set tblproperties("orc.bloom.filter.fpp"="0.01");
-insert overwrite table orc_ppd select * from staging;
+insert overwrite table orc_ppd_n0 select * from staging_n4;
-select * from orc_ppd limit 1;
+select * from orc_ppd_n0 limit 1;
CREATE TABLE orc_ppd_part(t tinyint,
si smallint,
@@ -55,6 +55,6 @@ CREATE TABLE orc_ppd_part(t tinyint,
bin binary)
PARTITIONED BY (ds string, hr int) STORED AS ORC tblproperties("orc.row.index.stride" = "1000", "orc.bloom.filter.columns"="*");
-insert overwrite table orc_ppd_part partition(ds = "2015", hr = 10) select * from staging;
+insert overwrite table orc_ppd_part partition(ds = "2015", hr = 10) select * from staging_n4;
select * from orc_ppd_part limit 1;
http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_int_type_promotion.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_int_type_promotion.q b/ql/src/test/queries/clientpositive/orc_int_type_promotion.q
index 0596e9f..853b317 100644
--- a/ql/src/test/queries/clientpositive/orc_int_type_promotion.q
+++ b/ql/src/test/queries/clientpositive/orc_int_type_promotion.q
@@ -1,7 +1,7 @@
set hive.vectorized.execution.enabled=false;
set hive.mapred.mode=nonstrict;
set hive.metastore.disallow.incompatible.col.type.changes=false;
-create table if not exists alltypes (
+create table if not exists alltypes_n0 (
bo boolean,
ti tinyint,
si smallint,
@@ -22,7 +22,7 @@ create table if not exists alltypes (
collection items terminated by ','
map keys terminated by ':' stored as textfile;
-create table if not exists alltypes_orc (
+create table if not exists alltypes_orc_n0 (
bo boolean,
ti tinyint,
si smallint,
@@ -41,26 +41,26 @@ create table if not exists alltypes_orc (
st struct<c1:int, c2:string>
) stored as orc;
-load data local inpath '../../data/files/alltypes2.txt' overwrite into table alltypes;
+load data local inpath '../../data/files/alltypes2.txt' overwrite into table alltypes_n0;
-insert overwrite table alltypes_orc select * from alltypes;
+insert overwrite table alltypes_orc_n0 select * from alltypes_n0;
-select * from alltypes_orc;
+select * from alltypes_orc_n0;
SET hive.exec.schema.evolution=true;
-alter table alltypes_orc change si si int;
-select * from alltypes_orc;
+alter table alltypes_orc_n0 change si si int;
+select * from alltypes_orc_n0;
-alter table alltypes_orc change si si bigint;
-alter table alltypes_orc change i i bigint;
-select * from alltypes_orc;
+alter table alltypes_orc_n0 change si si bigint;
+alter table alltypes_orc_n0 change i i bigint;
+select * from alltypes_orc_n0;
set hive.vectorized.execution.enabled=true;
set hive.fetch.task.conversion=none;
-explain select ti, si, i, bi from alltypes_orc;
-select ti, si, i, bi from alltypes_orc;
+explain select ti, si, i, bi from alltypes_orc_n0;
+select ti, si, i, bi from alltypes_orc_n0;
SET hive.exec.schema.evolution=false;