You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by gu...@apache.org on 2013/11/05 08:01:58 UTC
svn commit: r1538880 [12/46] - in /hive/branches/tez: ./ ant/
ant/src/org/apache/hadoop/hive/ant/ beeline/
beeline/src/java/org/apache/hive/beeline/ beeline/src/main/
beeline/src/test/org/apache/hive/beeline/src/test/ cli/ common/
common/src/java/conf/...
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/progress_1.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/progress_1.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/progress_1.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/progress_1.q Tue Nov 5 07:01:32 2013
@@ -2,7 +2,7 @@ set hive.heartbeat.interval=5;
CREATE TABLE PROGRESS_1(key int, value string) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/kv6.txt' INTO TABLE PROGRESS_1;
+LOAD DATA LOCAL INPATH '../../data/files/kv6.txt' INTO TABLE PROGRESS_1;
select count(1) from PROGRESS_1 t1 join PROGRESS_1 t2 on t1.key=t2.key;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/ptf.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/ptf.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/ptf.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/ptf.q Tue Nov 5 07:01:32 2013
@@ -13,7 +13,7 @@ CREATE TABLE part(
p_comment STRING
);
-LOAD DATA LOCAL INPATH '../data/files/part_tiny.txt' overwrite into table part;
+LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part;
--1. test1
select p_mfgr, p_name, p_size,
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/ptf_decimal.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/ptf_decimal.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/ptf_decimal.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/ptf_decimal.q Tue Nov 5 07:01:32 2013
@@ -13,7 +13,7 @@ CREATE TABLE part(
p_comment STRING
);
-LOAD DATA LOCAL INPATH '../data/files/part_tiny.txt' overwrite into table part;
+LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part;
-- 1. aggregate functions with decimal type
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/ptf_general_queries.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/ptf_general_queries.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/ptf_general_queries.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/ptf_general_queries.q Tue Nov 5 07:01:32 2013
@@ -13,7 +13,7 @@ CREATE TABLE part(
p_comment STRING
);
-LOAD DATA LOCAL INPATH '../data/files/part_tiny.txt' overwrite into table part;
+LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part;
-- 1. testNoPTFNoWindowing
select p_mfgr, p_name, p_size
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/ptf_matchpath.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/ptf_matchpath.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/ptf_matchpath.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/ptf_matchpath.q Tue Nov 5 07:01:32 2013
@@ -10,7 +10,7 @@ ARR_DELAY float,
FL_NUM string
);
-LOAD DATA LOCAL INPATH '../data/files/flights_tiny.txt' OVERWRITE INTO TABLE flights_tiny;
+LOAD DATA LOCAL INPATH '../../data/files/flights_tiny.txt' OVERWRITE INTO TABLE flights_tiny;
-- 1. basic Matchpath test
select origin_city_name, fl_num, year, month, day_of_month, sz, tpath
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/ptf_rcfile.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/ptf_rcfile.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/ptf_rcfile.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/ptf_rcfile.q Tue Nov 5 07:01:32 2013
@@ -12,7 +12,7 @@ CREATE TABLE part_rc(
p_comment STRING
) STORED AS RCFILE ;
-LOAD DATA LOCAL INPATH '../data/files/part.rc' overwrite into table part_rc;
+LOAD DATA LOCAL INPATH '../../data/files/part.rc' overwrite into table part_rc;
-- testWindowingPTFWithPartRC
select p_mfgr, p_name, p_size,
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/ptf_register_tblfn.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/ptf_register_tblfn.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/ptf_register_tblfn.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/ptf_register_tblfn.q Tue Nov 5 07:01:32 2013
@@ -10,7 +10,7 @@ ARR_DELAY float,
FL_NUM string
);
-LOAD DATA LOCAL INPATH '../data/files/flights_tiny.txt' OVERWRITE INTO TABLE flights_tiny;
+LOAD DATA LOCAL INPATH '../../data/files/flights_tiny.txt' OVERWRITE INTO TABLE flights_tiny;
create temporary function matchpathtest as 'org.apache.hadoop.hive.ql.udf.ptf.MatchPath$MatchPathResolver';
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/ptf_seqfile.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/ptf_seqfile.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/ptf_seqfile.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/ptf_seqfile.q Tue Nov 5 07:01:32 2013
@@ -12,7 +12,7 @@ CREATE TABLE part_seq(
p_comment STRING
) STORED AS SEQUENCEFILE ;
-LOAD DATA LOCAL INPATH '../data/files/part.seq' overwrite into table part_seq;
+LOAD DATA LOCAL INPATH '../../data/files/part.seq' overwrite into table part_seq;
-- testWindowingPTFWithPartSeqFile
select p_mfgr, p_name, p_size,
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/ql_rewrite_gbtoidx.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/ql_rewrite_gbtoidx.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/ql_rewrite_gbtoidx.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/ql_rewrite_gbtoidx.q Tue Nov 5 07:01:32 2013
@@ -19,7 +19,7 @@ CREATE TABLE lineitem (L_ORDERKEY I
ROW FORMAT DELIMITED
FIELDS TERMINATED BY '|';
-LOAD DATA LOCAL INPATH '../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem;
+LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem;
CREATE INDEX lineitem_lshipdate_idx ON TABLE lineitem(l_shipdate) AS 'org.apache.hadoop.hive.ql.index.AggregateIndexHandler' WITH DEFERRED REBUILD IDXPROPERTIES("AGGREGATES"="count(l_shipdate)");
ALTER INDEX lineitem_lshipdate_idx ON lineitem REBUILD;
@@ -156,7 +156,7 @@ DROP INDEX tbl_part_index on tblpart;
DROP TABLE tblpart;
CREATE TABLE tbl(key int, value int) ROW FORMAT DELIMITED FIELDS TERMINATED BY '|';
-LOAD DATA LOCAL INPATH '../data/files/tbl.txt' OVERWRITE INTO TABLE tbl;
+LOAD DATA LOCAL INPATH '../../data/files/tbl.txt' OVERWRITE INTO TABLE tbl;
CREATE INDEX tbl_key_idx ON TABLE tbl(key) AS 'org.apache.hadoop.hive.ql.index.AggregateIndexHandler' WITH DEFERRED REBUILD IDXPROPERTIES("AGGREGATES"="count(key)");
ALTER INDEX tbl_key_idx ON tbl REBUILD;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/quote2.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/quote2.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/quote2.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/quote2.q Tue Nov 5 07:01:32 2013
@@ -1,3 +1,5 @@
+set hive.fetch.task.conversion=more;
+
EXPLAIN
SELECT
'abc', "abc",
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/rcfile_bigdata.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/rcfile_bigdata.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/rcfile_bigdata.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/rcfile_bigdata.q Tue Nov 5 07:01:32 2013
@@ -1,7 +1,7 @@
set hive.map.aggr.hash.percentmemory = 0.3;
set hive.mapred.local.mem = 256;
-add file ../data/scripts/dumpdata_script.py;
+add file ../../data/scripts/dumpdata_script.py;
CREATE table columnTable_Bigdata (key STRING, value STRING)
ROW FORMAT SERDE
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/remote_script.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/remote_script.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/remote_script.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/remote_script.q Tue Nov 5 07:01:32 2013
@@ -1,4 +1,4 @@
-dfs -put ../data/scripts/newline.py /newline.py;
+dfs -put ../../data/scripts/newline.py /newline.py;
add file hdfs:///newline.py;
set hive.transform.escape.input=true;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/repair.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/repair.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/repair.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/repair.q Tue Nov 5 07:01:32 2013
@@ -1,10 +1,12 @@
+DROP TABLE IF EXISTS repairtable;
+
CREATE TABLE repairtable(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
MSCK TABLE repairtable;
-dfs ${system:test.dfs.mkdir} ../build/ql/test/data/warehouse/repairtable/p1=a/p2=a;
-dfs ${system:test.dfs.mkdir} ../build/ql/test/data/warehouse/repairtable/p1=b/p2=a;
-dfs -touchz ../build/ql/test/data/warehouse/repairtable/p1=b/p2=a/datafile;
+dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=a/p2=a;
+dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=b/p2=a;
+dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=b/p2=a/datafile;
MSCK TABLE repairtable;
@@ -12,4 +14,4 @@ MSCK REPAIR TABLE repairtable;
MSCK TABLE repairtable;
-
+DROP TABLE repairtable;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/scriptfile1.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/scriptfile1.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/scriptfile1.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/scriptfile1.q Tue Nov 5 07:01:32 2013
@@ -3,7 +3,7 @@ set hive.input.format=org.apache.hadoop.
-- EXCLUDE_OS_WINDOWS
CREATE TABLE dest1(key INT, value STRING);
-ADD FILE src/test/scripts/testgrep;
+ADD FILE ../../ql/src/test/scripts/testgrep;
FROM (
FROM src
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/serde_regex.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/serde_regex.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/serde_regex.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/serde_regex.q Tue Nov 5 07:01:32 2013
@@ -31,8 +31,8 @@ WITH SERDEPROPERTIES (
)
STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH "../data/files/apache.access.log" INTO TABLE serde_regex;
-LOAD DATA LOCAL INPATH "../data/files/apache.access.2.log" INTO TABLE serde_regex;
+LOAD DATA LOCAL INPATH "../../data/files/apache.access.log" INTO TABLE serde_regex;
+LOAD DATA LOCAL INPATH "../../data/files/apache.access.2.log" INTO TABLE serde_regex;
SELECT * FROM serde_regex ORDER BY time;
@@ -59,7 +59,7 @@ WITH SERDEPROPERTIES (
)
STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH "../data/files/kv7.txt" INTO TABLE serde_regex1;
+LOAD DATA LOCAL INPATH "../../data/files/kv7.txt" INTO TABLE serde_regex1;
SELECT key, value FROM serde_regex1 ORDER BY key, value;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/set_processor_namespaces.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/set_processor_namespaces.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/set_processor_namespaces.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/set_processor_namespaces.q Tue Nov 5 07:01:32 2013
@@ -24,7 +24,7 @@ set b=a;
set c=${hiveconf:${hiveconf:b}};
set c;
-set jar=${system:build.ivy.lib.dir}/default/derby-${system:derby.version}.jar;
+set jar=${system:maven.local.repository}/org/apache/derby/derby/${system:derby.version}/derby-${system:derby.version}.jar;
add file ${hiveconf:jar};
delete file ${hiveconf:jar};
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin.q Tue Nov 5 07:01:32 2013
@@ -13,10 +13,10 @@ CREATE TABLE T3(key STRING, val STRING)
CREATE TABLE T4(key STRING, val STRING) STORED AS TEXTFILE;
CREATE TABLE dest_j1(key INT, value STRING) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
-LOAD DATA LOCAL INPATH '../data/files/T3.txt' INTO TABLE T3;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T4;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T4;
EXPLAIN
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin_union_remove_1.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin_union_remove_1.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin_union_remove_1.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin_union_remove_1.q Tue Nov 5 07:01:32 2013
@@ -20,12 +20,12 @@ set mapred.input.dir.recursive=true;
CREATE TABLE T1(key STRING, val STRING)
SKEWED BY (key) ON ((2)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
CREATE TABLE T2(key STRING, val STRING)
SKEWED BY (key) ON ((3)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
-- a simple join query with skew on both the tables on the join key
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin_union_remove_2.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin_union_remove_2.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin_union_remove_2.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin_union_remove_2.q Tue Nov 5 07:01:32 2013
@@ -12,16 +12,16 @@ set mapred.input.dir.recursive=true;
CREATE TABLE T1(key STRING, val STRING)
SKEWED BY (key) ON ((2), (8)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
CREATE TABLE T2(key STRING, val STRING)
SKEWED BY (key) ON ((3), (8)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
CREATE TABLE T3(key STRING, val STRING) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T3.txt' INTO TABLE T3;
+LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3;
-- This is to test the union->selectstar->filesink and skewjoin optimization
-- Union of 3 map-reduce subqueries is performed for the skew join
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt1.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt1.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt1.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt1.q Tue Nov 5 07:01:32 2013
@@ -5,12 +5,12 @@ set hive.optimize.skewjoin.compiletime =
CREATE TABLE T1(key STRING, val STRING)
SKEWED BY (key) ON ((2)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
CREATE TABLE T2(key STRING, val STRING)
SKEWED BY (key) ON ((3)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
-- a simple join query with skew on both the tables on the join key
-- adding a order by at the end to make the results deterministic
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt10.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt10.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt10.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt10.q Tue Nov 5 07:01:32 2013
@@ -4,7 +4,7 @@ set hive.optimize.skewjoin.compiletime =
CREATE TABLE T1(key STRING, value STRING) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
drop table array_valued_T1;
create table array_valued_T1 (key string, value array<string>) SKEWED BY (key) ON ((8));
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt11.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt11.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt11.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt11.q Tue Nov 5 07:01:32 2013
@@ -5,11 +5,11 @@ set hive.optimize.skewjoin.compiletime =
CREATE TABLE T1(key STRING, val STRING)
SKEWED BY (key) ON ((2)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
-- This test is to verify the skew join compile optimization when the join is followed
-- by a union. Both sides of a union consist of a join, which should have used
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt12.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt12.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt12.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt12.q Tue Nov 5 07:01:32 2013
@@ -5,12 +5,12 @@ set hive.optimize.skewjoin.compiletime =
CREATE TABLE T1(key STRING, val STRING)
SKEWED BY (key, val) ON ((2, 12), (8, 18)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
CREATE TABLE T2(key STRING, val STRING)
SKEWED BY (key, val) ON ((3, 13), (8, 18)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
-- Both the join tables are skewed by 2 keys, and one of the skewed values
-- is common to both the tables. The join key matches the skewed key set.
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt13.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt13.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt13.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt13.q Tue Nov 5 07:01:32 2013
@@ -4,16 +4,16 @@ set hive.optimize.skewjoin.compiletime =
CREATE TABLE T1(key STRING, val STRING) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
CREATE TABLE T3(key STRING, val STRING)
SKEWED BY (val) ON ((12)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T3.txt' INTO TABLE T3;
+LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3;
-- This test is for skewed join compile time optimization for more than 2 tables.
-- The join key for table 3 is different from the join key used for joining
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt14.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt14.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt14.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt14.q Tue Nov 5 07:01:32 2013
@@ -5,16 +5,16 @@ set hive.optimize.skewjoin.compiletime =
CREATE TABLE T1(key STRING, val STRING)
SKEWED BY (key) ON ((2)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
CREATE TABLE T3(key STRING, val STRING)
SKEWED BY (val) ON ((12)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T3.txt' INTO TABLE T3;
+LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3;
-- This test is for skewed join compile time optimization for more than 2 tables.
-- The join key for table 3 is different from the join key used for joining
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt15.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt15.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt15.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt15.q Tue Nov 5 07:01:32 2013
@@ -4,7 +4,7 @@ set hive.optimize.skewjoin.compiletime =
CREATE TABLE tmpT1(key STRING, val STRING) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE tmpT1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE tmpT1;
-- testing skew on other data types - int
CREATE TABLE T1(key INT, val STRING) SKEWED BY (key) ON ((2));
@@ -12,7 +12,7 @@ INSERT OVERWRITE TABLE T1 SELECT key, va
CREATE TABLE tmpT2(key STRING, val STRING) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE tmpT2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE tmpT2;
CREATE TABLE T2(key INT, val STRING) SKEWED BY (key) ON ((3));
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt16.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt16.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt16.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt16.q Tue Nov 5 07:01:32 2013
@@ -5,12 +5,12 @@ set hive.optimize.skewjoin.compiletime =
CREATE TABLE T1(key STRING, val STRING)
SKEWED BY (key, val) ON ((2, 12)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
CREATE TABLE T2(key STRING, val STRING)
SKEWED BY (key) ON ((3)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
-- One of the tables is skewed by 2 columns, and the other table is
-- skewed by one column. Ths join is performed on the both the columns
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt17.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt17.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt17.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt17.q Tue Nov 5 07:01:32 2013
@@ -5,12 +5,12 @@ set hive.optimize.skewjoin.compiletime =
CREATE TABLE T1(key STRING, val STRING)
SKEWED BY (key, val) ON ((2, 12)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
CREATE TABLE T2(key STRING, val STRING)
SKEWED BY (key) ON ((2)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
-- One of the tables is skewed by 2 columns, and the other table is
-- skewed by one column. Ths join is performed on the first skewed column
@@ -31,12 +31,12 @@ DROP TABLE T2;
CREATE TABLE T1(key STRING, val STRING)
SKEWED BY (key, val) ON ((2, 12)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
CREATE TABLE T2(key STRING, val STRING)
SKEWED BY (key) ON ((2)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
-- One of the tables is skewed by 2 columns, and the other table is
-- skewed by one column. Ths join is performed on the both the columns
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt18.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt18.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt18.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt18.q Tue Nov 5 07:01:32 2013
@@ -4,7 +4,7 @@ set hive.optimize.skewjoin.compiletime =
CREATE TABLE tmpT1(key STRING, val STRING) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE tmpT1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE tmpT1;
-- testing skew on other data types - int
CREATE TABLE T1(key INT, val STRING) SKEWED BY (key) ON ((2));
@@ -16,7 +16,7 @@ INSERT OVERWRITE TABLE T1 SELECT key, va
CREATE TABLE T2(key STRING, val STRING)
SKEWED BY (key) ON ((3)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
-- Once HIVE-3445 is fixed, the compile time skew join optimization would be
-- applicable here. Till the above jira is fixed, it would be performed as a
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt19.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt19.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt19.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt19.q Tue Nov 5 07:01:32 2013
@@ -6,11 +6,11 @@ CREATE TABLE T1(key STRING, val STRING)
CLUSTERED BY (key) INTO 4 BUCKETS
SKEWED BY (key) ON ((2)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
-- add a test where the skewed key is also the bucketized key
-- it should not matter, and the compile time skewed join
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt2.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt2.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt2.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt2.q Tue Nov 5 07:01:32 2013
@@ -5,12 +5,12 @@ set hive.optimize.skewjoin.compiletime =
CREATE TABLE T1(key STRING, val STRING)
SKEWED BY (key) ON ((2), (7)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
CREATE TABLE T2(key STRING, val STRING)
SKEWED BY (key) ON ((3), (8)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
-- a simple query with skew on both the tables on the join key
-- multiple skew values are present for the skewed keys
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt20.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt20.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt20.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt20.q Tue Nov 5 07:01:32 2013
@@ -6,11 +6,11 @@ CREATE TABLE T1(key STRING, val STRING)
CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS
SKEWED BY (key) ON ((2)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
-- add a test where the skewed key is also the bucketized/sorted key
-- it should not matter, and the compile time skewed join
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt3.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt3.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt3.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt3.q Tue Nov 5 07:01:32 2013
@@ -5,12 +5,12 @@ set hive.optimize.skewjoin.compiletime =
CREATE TABLE T1(key STRING, val STRING)
SKEWED BY (key) ON ((2), (8)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
CREATE TABLE T2(key STRING, val STRING)
SKEWED BY (key) ON ((3), (8)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
-- a simple query with skew on both the tables. One of the skewed
-- value is common to both the tables. The skewed value should not be
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt4.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt4.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt4.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt4.q Tue Nov 5 07:01:32 2013
@@ -5,11 +5,11 @@ set hive.optimize.skewjoin.compiletime =
CREATE TABLE T1(key STRING, val STRING)
SKEWED BY (key) ON ((2)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
-- only of the tables of the join (the left table of the join) is skewed
-- the skewed filter would still be applied to both the tables
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt5.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt5.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt5.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt5.q Tue Nov 5 07:01:32 2013
@@ -5,12 +5,12 @@ set hive.optimize.skewjoin.compiletime =
CREATE TABLE T1(key STRING, val STRING)
SKEWED BY (key, val) ON ((2, 12)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
CREATE TABLE T2(key STRING, val STRING)
SKEWED BY (key) ON ((3)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
-- One of the tables is skewed by 2 columns, and the other table is
-- skewed by one column. Ths join is performed on the first skewed column
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt6.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt6.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt6.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt6.q Tue Nov 5 07:01:32 2013
@@ -5,12 +5,12 @@ set hive.optimize.skewjoin.compiletime =
CREATE TABLE T1(key STRING, val STRING)
SKEWED BY (key, val) ON ((2, 12), (8, 18)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
CREATE TABLE T2(key STRING, val STRING)
SKEWED BY (key, val) ON ((3, 13), (8, 18)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
-- Both the join tables are skewed by 2 keys, and one of the skewed values
-- is common to both the tables. The join key is a subset of the skewed key set:
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt7.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt7.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt7.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt7.q Tue Nov 5 07:01:32 2013
@@ -5,16 +5,16 @@ set hive.optimize.skewjoin.compiletime =
CREATE TABLE T1(key STRING, val STRING)
SKEWED BY (key) ON ((2), (8)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
CREATE TABLE T2(key STRING, val STRING)
SKEWED BY (key) ON ((3), (8)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
CREATE TABLE T3(key STRING, val STRING) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T3.txt' INTO TABLE T3;
+LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3;
-- This test is for validating skewed join compile time optimization for more than
-- 2 tables. The join key is the same, and so a 3-way join would be performed.
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt8.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt8.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt8.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt8.q Tue Nov 5 07:01:32 2013
@@ -4,16 +4,16 @@ set hive.optimize.skewjoin.compiletime =
CREATE TABLE T1(key STRING, val STRING) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
CREATE TABLE T2(key STRING, val STRING)
SKEWED BY (key) ON ((3), (8)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
CREATE TABLE T3(key STRING, val STRING) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T3.txt' INTO TABLE T3;
+LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3;
-- This test is for validating skewed join compile time optimization for more than
-- 2 tables. The join key is the same, and so a 3-way join would be performed.
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt9.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt9.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt9.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt9.q Tue Nov 5 07:01:32 2013
@@ -5,11 +5,11 @@ set hive.optimize.skewjoin.compiletime =
CREATE TABLE T1(key STRING, val STRING)
SKEWED BY (key) ON ((2)) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
-- no skew join compile time optimization would be performed if one of the
-- join sources is a sub-query consisting of a union all
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_1.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_1.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_1.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_1.q Tue Nov 5 07:01:32 2013
@@ -6,9 +6,9 @@ create table smb_bucket_1(key int, value
create table smb_bucket_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
create table smb_bucket_3(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
-load data local inpath '../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
-load data local inpath '../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
-load data local inpath '../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
+load data local inpath '../../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
+load data local inpath '../../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
+load data local inpath '../../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
set hive.optimize.bucketmapjoin = true;
set hive.optimize.bucketmapjoin.sortedmerge = true;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_10.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_10.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_10.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_10.q Tue Nov 5 07:01:32 2013
@@ -6,11 +6,11 @@ alter table tmp_smb_bucket_10 add partit
-- add dummy files to make sure that the number of files in each partition is same as number of buckets
-load data local inpath '../data/files/smbbucket_1.rc' INTO TABLE tmp_smb_bucket_10 partition(ds='1');
-load data local inpath '../data/files/smbbucket_2.rc' INTO TABLE tmp_smb_bucket_10 partition(ds='1');
+load data local inpath '../../data/files/smbbucket_1.rc' INTO TABLE tmp_smb_bucket_10 partition(ds='1');
+load data local inpath '../../data/files/smbbucket_2.rc' INTO TABLE tmp_smb_bucket_10 partition(ds='1');
-load data local inpath '../data/files/smbbucket_1.rc' INTO TABLE tmp_smb_bucket_10 partition(ds='2');
-load data local inpath '../data/files/smbbucket_2.rc' INTO TABLE tmp_smb_bucket_10 partition(ds='2');
+load data local inpath '../../data/files/smbbucket_1.rc' INTO TABLE tmp_smb_bucket_10 partition(ds='2');
+load data local inpath '../../data/files/smbbucket_2.rc' INTO TABLE tmp_smb_bucket_10 partition(ds='2');
set hive.optimize.bucketmapjoin = true;
set hive.optimize.bucketmapjoin.sortedmerge = true;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_2.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_2.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_2.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_2.q Tue Nov 5 07:01:32 2013
@@ -6,9 +6,9 @@ create table smb_bucket_1(key int, value
create table smb_bucket_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
create table smb_bucket_3(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
-load data local inpath '../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
-load data local inpath '../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
-load data local inpath '../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
+load data local inpath '../../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
+load data local inpath '../../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
+load data local inpath '../../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
set hive.optimize.bucketmapjoin = true;
set hive.optimize.bucketmapjoin.sortedmerge = true;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_25.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_25.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_25.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_25.q Tue Nov 5 07:01:32 2013
@@ -10,9 +10,9 @@ create table smb_bucket_1(key int, value
create table smb_bucket_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
create table smb_bucket_3(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
-load data local inpath '../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
-load data local inpath '../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
-load data local inpath '../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
+load data local inpath '../../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
+load data local inpath '../../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
+load data local inpath '../../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
explain
select * from (select a.key from smb_bucket_1 a join smb_bucket_2 b on (a.key = b.key) where a.key = 5) t1 left outer join (select c.key from smb_bucket_2 c join smb_bucket_3 d on (c.key = d.key) where c.key=5) t2 on (t1.key=t2.key) where t2.key=5;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_3.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_3.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_3.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_3.q Tue Nov 5 07:01:32 2013
@@ -6,9 +6,9 @@ create table smb_bucket_1(key int, value
create table smb_bucket_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
create table smb_bucket_3(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
-load data local inpath '../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
-load data local inpath '../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
-load data local inpath '../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
+load data local inpath '../../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
+load data local inpath '../../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
+load data local inpath '../../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
set hive.optimize.bucketmapjoin = true;
set hive.optimize.bucketmapjoin.sortedmerge = true;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_4.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_4.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_4.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_4.q Tue Nov 5 07:01:32 2013
@@ -6,9 +6,9 @@ create table smb_bucket_1(key int, value
create table smb_bucket_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
create table smb_bucket_3(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
-load data local inpath '../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
-load data local inpath '../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
-load data local inpath '../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
+load data local inpath '../../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
+load data local inpath '../../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
+load data local inpath '../../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
set hive.optimize.bucketmapjoin = true;
set hive.optimize.bucketmapjoin.sortedmerge = true;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_5.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_5.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_5.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_5.q Tue Nov 5 07:01:32 2013
@@ -6,9 +6,9 @@ create table smb_bucket_1(key int, value
create table smb_bucket_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
create table smb_bucket_3(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
-load data local inpath '../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
-load data local inpath '../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
-load data local inpath '../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
+load data local inpath '../../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
+load data local inpath '../../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
+load data local inpath '../../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
set hive.optimize.bucketmapjoin = true;
set hive.optimize.bucketmapjoin.sortedmerge = true;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_7.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_7.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_7.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_7.q Tue Nov 5 07:01:32 2013
@@ -15,8 +15,8 @@ create table smb_join_results(k1 int, v1
create table smb_join_results_empty_bigtable(k1 int, v1 string, k2 int, v2 string);
create table normal_join_results(k1 int, v1 string, k2 int, v2 string);
-load data local inpath '../data/files/empty1.txt' into table smb_bucket4_1;
-load data local inpath '../data/files/empty2.txt' into table smb_bucket4_1;
+load data local inpath '../../data/files/empty1.txt' into table smb_bucket4_1;
+load data local inpath '../../data/files/empty2.txt' into table smb_bucket4_1;
insert overwrite table smb_bucket4_2
select * from src;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_8.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_8.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_8.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_8.q Tue Nov 5 07:01:32 2013
@@ -5,7 +5,7 @@ set hive.exec.reducers.max = 1;
create table smb_bucket_input (key int, value string) stored as rcfile;
-load data local inpath '../data/files/smb_bucket_input.rc' into table smb_bucket_input;
+load data local inpath '../../data/files/smb_bucket_input.rc' into table smb_bucket_input;
CREATE TABLE smb_bucket4_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/source.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/source.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/source.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/source.q Tue Nov 5 07:01:32 2013
@@ -1 +1 @@
-source ../data/files/source.txt;
+source ../../data/files/source.txt;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/split.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/split.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/split.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/split.q Tue Nov 5 07:01:32 2013
@@ -2,7 +2,7 @@ DROP TABLE tmp_jo_tab_test;
CREATE table tmp_jo_tab_test (message_line STRING)
STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/input.txt'
+LOAD DATA LOCAL INPATH '../../data/files/input.txt'
OVERWRITE INTO TABLE tmp_jo_tab_test;
select size(split(message_line, '\t')) from tmp_jo_tab_test;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/stats1.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/stats1.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/stats1.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/stats1.q Tue Nov 5 07:01:32 2013
@@ -26,5 +26,5 @@ DESCRIBE FORMATTED tmptable;
-- Load a file into a existing table
-- Some stats (numFiles, totalSize) should be updated correctly
-- Some other stats (numRows, rawDataSize) should be cleared
-load data local inpath '../data/files/srcbucket20.txt' INTO TABLE tmptable;
+load data local inpath '../../data/files/srcbucket20.txt' INTO TABLE tmptable;
DESCRIBE FORMATTED tmptable;
\ No newline at end of file
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/stats11.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/stats11.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/stats11.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/stats11.q Tue Nov 5 07:01:32 2013
@@ -2,25 +2,25 @@ set datanucleus.cache.collections=false;
set hive.stats.autogather=true;
CREATE TABLE srcbucket_mapjoin(key int, value string) CLUSTERED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE;
-load data local inpath '../data/files/srcbucket20.txt' INTO TABLE srcbucket_mapjoin;
-load data local inpath '../data/files/srcbucket21.txt' INTO TABLE srcbucket_mapjoin;
+load data local inpath '../../data/files/srcbucket20.txt' INTO TABLE srcbucket_mapjoin;
+load data local inpath '../../data/files/srcbucket21.txt' INTO TABLE srcbucket_mapjoin;
CREATE TABLE srcbucket_mapjoin_part (key int, value string) partitioned by (ds string) CLUSTERED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE;
explain
-load data local inpath '../data/files/srcbucket20.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
-load data local inpath '../data/files/srcbucket20.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
+load data local inpath '../../data/files/srcbucket20.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
+load data local inpath '../../data/files/srcbucket20.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
desc formatted srcbucket_mapjoin_part partition(ds='2008-04-08');
-load data local inpath '../data/files/srcbucket21.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
+load data local inpath '../../data/files/srcbucket21.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
desc formatted srcbucket_mapjoin_part partition(ds='2008-04-08');
-load data local inpath '../data/files/srcbucket22.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
+load data local inpath '../../data/files/srcbucket22.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
desc formatted srcbucket_mapjoin_part partition(ds='2008-04-08');
-load data local inpath '../data/files/srcbucket23.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
+load data local inpath '../../data/files/srcbucket23.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
desc formatted srcbucket_mapjoin_part partition(ds='2008-04-08');
CREATE TABLE srcbucket_mapjoin_part_2 (key int, value string) partitioned by (ds string) CLUSTERED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE;
-load data local inpath '../data/files/srcbucket22.txt' INTO TABLE srcbucket_mapjoin_part_2 partition(ds='2008-04-08');
-load data local inpath '../data/files/srcbucket23.txt' INTO TABLE srcbucket_mapjoin_part_2 partition(ds='2008-04-08');
+load data local inpath '../../data/files/srcbucket22.txt' INTO TABLE srcbucket_mapjoin_part_2 partition(ds='2008-04-08');
+load data local inpath '../../data/files/srcbucket23.txt' INTO TABLE srcbucket_mapjoin_part_2 partition(ds='2008-04-08');
create table bucketmapjoin_hash_result_1 (key bigint , value1 bigint, value2 bigint);
create table bucketmapjoin_hash_result_2 (key bigint , value1 bigint, value2 bigint);
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/stats18.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/stats18.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/stats18.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/stats18.q Tue Nov 5 07:01:32 2013
@@ -13,7 +13,7 @@ insert overwrite table stats_part partit
-- Some other stats (numRows, rawDataSize) should be cleared
desc formatted stats_part partition (ds='2010-04-08', hr='13');
-load data local inpath '../data/files/srcbucket20.txt' INTO TABLE stats_part partition (ds='2010-04-08', hr='13');
+load data local inpath '../../data/files/srcbucket20.txt' INTO TABLE stats_part partition (ds='2010-04-08', hr='13');
desc formatted stats_part partition (ds='2010-04-08', hr='13');
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/stats3.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/stats3.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/stats3.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/stats3.q Tue Nov 5 07:01:32 2013
@@ -5,9 +5,9 @@ drop table hive_test_dst;
create table hive_test_src ( col1 string ) stored as textfile ;
explain extended
-load data local inpath '../data/files/test.dat' overwrite into table hive_test_src ;
+load data local inpath '../../data/files/test.dat' overwrite into table hive_test_src ;
-load data local inpath '../data/files/test.dat' overwrite into table hive_test_src ;
+load data local inpath '../../data/files/test.dat' overwrite into table hive_test_src ;
desc formatted hive_test_src;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/stats_noscan_2.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/stats_noscan_2.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/stats_noscan_2.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/stats_noscan_2.q Tue Nov 5 07:01:32 2013
@@ -1,7 +1,7 @@
-- test analyze table compute statistiscs [noscan] on external table
-- 1 test table
-CREATE EXTERNAL TABLE anaylyze_external (a INT) LOCATION '${system:test.src.data.dir}/files/ext_test';
+CREATE EXTERNAL TABLE anaylyze_external (a INT) LOCATION '${system:hive.root}/data/files/ext_test';
SELECT * FROM anaylyze_external;
analyze table anaylyze_external compute statistics;
describe formatted anaylyze_external;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/str_to_map.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/str_to_map.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/str_to_map.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/str_to_map.q Tue Nov 5 07:01:32 2013
@@ -1,3 +1,5 @@
+set hive.fetch.task.conversion=more;
+
desc function str_to_map;
desc function extended str_to_map;
@@ -19,7 +21,7 @@ limit 3;
drop table tbl_s2m;
-create table tbl_s2m as select 'ABC=CC_333=444' as t from src limit 3;
+create table tbl_s2m as select 'ABC=CC_333=444' as t from src tablesample (3 rows);
select str_to_map(t,'_','=')['333'] from tbl_s2m;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/subq.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/subq.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/subq.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/subq.q Tue Nov 5 07:01:32 2013
@@ -2,12 +2,12 @@ EXPLAIN
FROM (
FROM src select src.* WHERE src.key < 100
) unioninput
-INSERT OVERWRITE DIRECTORY '../build/ql/test/data/warehouse/union.out' SELECT unioninput.*;
+INSERT OVERWRITE DIRECTORY 'target/warehouse/union.out' SELECT unioninput.*;
FROM (
FROM src select src.* WHERE src.key < 100
) unioninput
-INSERT OVERWRITE DIRECTORY '../build/ql/test/data/warehouse/union.out' SELECT unioninput.*;
+INSERT OVERWRITE DIRECTORY 'target/warehouse/union.out' SELECT unioninput.*;
-dfs -cat ../build/ql/test/data/warehouse/union.out/*;
+dfs -cat ${system:test.warehouse.dir}/union.out/*;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/subquery_in.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/subquery_in.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/subquery_in.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/subquery_in.q Tue Nov 5 07:01:32 2013
@@ -13,7 +13,7 @@ CREATE TABLE part(
p_comment STRING
);
-LOAD DATA LOCAL INPATH '../data/files/part_tiny.txt' overwrite into table part;
+LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part;
DROP TABLE lineitem;
CREATE TABLE lineitem (L_ORDERKEY INT,
@@ -35,7 +35,7 @@ CREATE TABLE lineitem (L_ORDERKEY I
ROW FORMAT DELIMITED
FIELDS TERMINATED BY '|';
-LOAD DATA LOCAL INPATH '../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem;
+LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem;
-- non agg, non corr
explain
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/subquery_notin.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/subquery_notin.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/subquery_notin.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/subquery_notin.q Tue Nov 5 07:01:32 2013
@@ -13,7 +13,7 @@ CREATE TABLE part(
p_comment STRING
);
-LOAD DATA LOCAL INPATH '../data/files/part_tiny.txt' overwrite into table part;
+LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part;
DROP TABLE lineitem;
CREATE TABLE lineitem (L_ORDERKEY INT,
@@ -35,7 +35,7 @@ CREATE TABLE lineitem (L_ORDERKEY I
ROW FORMAT DELIMITED
FIELDS TERMINATED BY '|';
-LOAD DATA LOCAL INPATH '../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem;
+LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem;
-- non agg, non corr
explain
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/symlink_text_input_format.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/symlink_text_input_format.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/symlink_text_input_format.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/symlink_text_input_format.q Tue Nov 5 07:01:32 2013
@@ -1,12 +1,12 @@
-
+DROP TABLE IF EXISTS symlink_text_input_format;
EXPLAIN
CREATE TABLE symlink_text_input_format (key STRING, value STRING) STORED AS INPUTFORMAT 'org.apache.hadoop.hive.ql.io.SymlinkTextInputFormat' OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat';
CREATE TABLE symlink_text_input_format (key STRING, value STRING) STORED AS INPUTFORMAT 'org.apache.hadoop.hive.ql.io.SymlinkTextInputFormat' OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat';
-dfs -cp ../data/files/symlink1.txt ../build/ql/test/data/warehouse/symlink_text_input_format/symlink1.txt;
-dfs -cp ../data/files/symlink2.txt ../build/ql/test/data/warehouse/symlink_text_input_format/symlink2.txt;
+dfs -cp ../../data/files/symlink1.txt ${system:test.warehouse.dir}/symlink_text_input_format/symlink1.txt;
+dfs -cp ../../data/files/symlink2.txt ${system:test.warehouse.dir}/symlink_text_input_format/symlink2.txt;
EXPLAIN SELECT * FROM symlink_text_input_format order by key, value;
@@ -20,5 +20,4 @@ EXPLAIN SELECT count(1) FROM symlink_tex
SELECT count(1) FROM symlink_text_input_format;
-
-
+DROP TABLE symlink_text_input_format;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/table_access_keys_stats.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/table_access_keys_stats.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/table_access_keys_stats.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/table_access_keys_stats.q Tue Nov 5 07:01:32 2013
@@ -4,7 +4,7 @@ SET hive.stats.collect.tablekeys=true;
-- This test is used for testing the TableAccessAnalyzer
CREATE TABLE T1(key STRING, val STRING) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;
CREATE TABLE T3(key STRING, val STRING) STORED AS TEXTFILE;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/test_boolean_whereclause.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/test_boolean_whereclause.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/test_boolean_whereclause.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/test_boolean_whereclause.q Tue Nov 5 07:01:32 2013
@@ -1,5 +1,5 @@
create table if not exists test_boolean(dummy tinyint);
-insert overwrite table test_boolean select 1 from src limit 1;
+insert overwrite table test_boolean select 1 from src tablesample (1 rows);
SELECT 1
FROM (
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_1.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_1.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_1.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_1.q Tue Nov 5 07:01:32 2013
@@ -1,10 +1,12 @@
+set hive.fetch.task.conversion=more;
+
drop table timestamp_1;
create table timestamp_1 (t timestamp);
alter table timestamp_1 set serde 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe';
-insert overwrite table timestamp_1
- select cast('2011-01-01 01:01:01' as timestamp) from src limit 1;
+insert overwrite table timestamp_1
+ select cast('2011-01-01 01:01:01' as timestamp) from src tablesample (1 rows);
select cast(t as boolean) from timestamp_1 limit 1;
select cast(t as tinyint) from timestamp_1 limit 1;
select cast(t as smallint) from timestamp_1 limit 1;
@@ -15,7 +17,7 @@ select cast(t as double) from timestamp_
select cast(t as string) from timestamp_1 limit 1;
insert overwrite table timestamp_1
- select '2011-01-01 01:01:01' from src limit 1;
+ select '2011-01-01 01:01:01' from src tablesample (1 rows);
select cast(t as boolean) from timestamp_1 limit 1;
select cast(t as tinyint) from timestamp_1 limit 1;
select cast(t as smallint) from timestamp_1 limit 1;
@@ -26,7 +28,7 @@ select cast(t as double) from timestamp_
select cast(t as string) from timestamp_1 limit 1;
insert overwrite table timestamp_1
- select '2011-01-01 01:01:01.1' from src limit 1;
+ select '2011-01-01 01:01:01.1' from src tablesample (1 rows);
select cast(t as boolean) from timestamp_1 limit 1;
select cast(t as tinyint) from timestamp_1 limit 1;
select cast(t as smallint) from timestamp_1 limit 1;
@@ -37,7 +39,7 @@ select cast(t as double) from timestamp_
select cast(t as string) from timestamp_1 limit 1;
insert overwrite table timestamp_1
- select '2011-01-01 01:01:01.0001' from src limit 1;
+ select '2011-01-01 01:01:01.0001' from src tablesample (1 rows);
select cast(t as boolean) from timestamp_1 limit 1;
select cast(t as tinyint) from timestamp_1 limit 1;
select cast(t as smallint) from timestamp_1 limit 1;
@@ -48,7 +50,7 @@ select cast(t as double) from timestamp_
select cast(t as string) from timestamp_1 limit 1;
insert overwrite table timestamp_1
- select '2011-01-01 01:01:01.000100000' from src limit 1;
+ select '2011-01-01 01:01:01.000100000' from src tablesample (1 rows);
select cast(t as boolean) from timestamp_1 limit 1;
select cast(t as tinyint) from timestamp_1 limit 1;
select cast(t as smallint) from timestamp_1 limit 1;
@@ -59,7 +61,7 @@ select cast(t as double) from timestamp_
select cast(t as string) from timestamp_1 limit 1;
insert overwrite table timestamp_1
- select '2011-01-01 01:01:01.001000011' from src limit 1;
+ select '2011-01-01 01:01:01.001000011' from src tablesample (1 rows);
select cast(t as boolean) from timestamp_1 limit 1;
select cast(t as tinyint) from timestamp_1 limit 1;
select cast(t as smallint) from timestamp_1 limit 1;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_2.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_2.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_2.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_2.q Tue Nov 5 07:01:32 2013
@@ -1,10 +1,12 @@
+set hive.fetch.task.conversion=more;
+
drop table timestamp_2;
create table timestamp_2 (t timestamp);
alter table timestamp_2 set serde 'org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe';
-insert overwrite table timestamp_2
- select cast('2011-01-01 01:01:01' as timestamp) from src limit 1;
+insert overwrite table timestamp_2
+ select cast('2011-01-01 01:01:01' as timestamp) from src tablesample (1 rows);
select cast(t as boolean) from timestamp_2 limit 1;
select cast(t as tinyint) from timestamp_2 limit 1;
select cast(t as smallint) from timestamp_2 limit 1;
@@ -15,7 +17,7 @@ select cast(t as double) from timestamp_
select cast(t as string) from timestamp_2 limit 1;
insert overwrite table timestamp_2
- select '2011-01-01 01:01:01' from src limit 1;
+ select '2011-01-01 01:01:01' from src tablesample (1 rows);
select cast(t as boolean) from timestamp_2 limit 1;
select cast(t as tinyint) from timestamp_2 limit 1;
select cast(t as smallint) from timestamp_2 limit 1;
@@ -26,7 +28,7 @@ select cast(t as double) from timestamp_
select cast(t as string) from timestamp_2 limit 1;
insert overwrite table timestamp_2
- select '2011-01-01 01:01:01.1' from src limit 1;
+ select '2011-01-01 01:01:01.1' from src tablesample (1 rows);
select cast(t as boolean) from timestamp_2 limit 1;
select cast(t as tinyint) from timestamp_2 limit 1;
select cast(t as smallint) from timestamp_2 limit 1;
@@ -37,7 +39,7 @@ select cast(t as double) from timestamp_
select cast(t as string) from timestamp_2 limit 1;
insert overwrite table timestamp_2
- select '2011-01-01 01:01:01.0001' from src limit 1;
+ select '2011-01-01 01:01:01.0001' from src tablesample (1 rows);
select cast(t as boolean) from timestamp_2 limit 1;
select cast(t as tinyint) from timestamp_2 limit 1;
select cast(t as smallint) from timestamp_2 limit 1;
@@ -48,7 +50,7 @@ select cast(t as double) from timestamp_
select cast(t as string) from timestamp_2 limit 1;
insert overwrite table timestamp_2
- select '2011-01-01 01:01:01.000100000' from src limit 1;
+ select '2011-01-01 01:01:01.000100000' from src tablesample (1 rows);
select cast(t as boolean) from timestamp_2 limit 1;
select cast(t as tinyint) from timestamp_2 limit 1;
select cast(t as smallint) from timestamp_2 limit 1;
@@ -59,7 +61,7 @@ select cast(t as double) from timestamp_
select cast(t as string) from timestamp_2 limit 1;
insert overwrite table timestamp_2
- select '2011-01-01 01:01:01.001000011' from src limit 1;
+ select '2011-01-01 01:01:01.001000011' from src tablesample (1 rows);
select cast(t as boolean) from timestamp_2 limit 1;
select cast(t as tinyint) from timestamp_2 limit 1;
select cast(t as smallint) from timestamp_2 limit 1;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_3.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_3.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_3.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_3.q Tue Nov 5 07:01:32 2013
@@ -1,10 +1,12 @@
+set hive.fetch.task.conversion=more;
+
drop table timestamp_3;
create table timestamp_3 (t timestamp);
alter table timestamp_3 set serde 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe';
-insert overwrite table timestamp_3
- select cast(cast('1.3041352164485E9' as double) as timestamp) from src limit 1;
+insert overwrite table timestamp_3
+ select cast(cast('1.3041352164485E9' as double) as timestamp) from src tablesample (1 rows);
select cast(t as boolean) from timestamp_3 limit 1;
select cast(t as tinyint) from timestamp_3 limit 1;
select cast(t as smallint) from timestamp_3 limit 1;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_comparison.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_comparison.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_comparison.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_comparison.q Tue Nov 5 07:01:32 2013
@@ -1,5 +1,6 @@
+set hive.fetch.task.conversion=more;
-select cast('2011-05-06 07:08:09' as timestamp) >
+select cast('2011-05-06 07:08:09' as timestamp) >
cast('2011-05-06 07:08:09' as timestamp) from src limit 1;
select cast('2011-05-06 07:08:09' as timestamp) <
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_lazy.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_lazy.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_lazy.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_lazy.q Tue Nov 5 07:01:32 2013
@@ -1,6 +1,6 @@
drop table timestamp_lazy;
create table timestamp_lazy (t timestamp, key string, value string);
-insert overwrite table timestamp_lazy select cast('2011-01-01 01:01:01' as timestamp), key, value from src limit 5;
+insert overwrite table timestamp_lazy select cast('2011-01-01 01:01:01' as timestamp), key, value from src tablesample (5 rows);
select t,key,value from timestamp_lazy ORDER BY key ASC, value ASC;
select t,key,value from timestamp_lazy distribute by t sort by key ASC, value ASC;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_null.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_null.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_null.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_null.q Tue Nov 5 07:01:32 2013
@@ -1,6 +1,6 @@
DROP TABLE IF EXISTS timestamp_null;
CREATE TABLE timestamp_null (t1 TIMESTAMP);
-LOAD DATA LOCAL INPATH '../data/files/test.dat' OVERWRITE INTO TABLE timestamp_null;
+LOAD DATA LOCAL INPATH '../../data/files/test.dat' OVERWRITE INTO TABLE timestamp_null;
SELECT * FROM timestamp_null LIMIT 1;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_udf.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_udf.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_udf.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_udf.q Tue Nov 5 07:01:32 2013
@@ -1,13 +1,15 @@
+set hive.fetch.task.conversion=more;
+
drop table timestamp_udf;
drop table timestamp_udf_string;
create table timestamp_udf (t timestamp);
create table timestamp_udf_string (t string);
-from src
+from (select * from src tablesample (1 rows)) s
insert overwrite table timestamp_udf
- select '2011-05-06 07:08:09.1234567' limit 1
+ select '2011-05-06 07:08:09.1234567'
insert overwrite table timestamp_udf_string
- select '2011-05-06 07:08:09.1234567' limit 1;
+ select '2011-05-06 07:08:09.1234567';
-- Test UDFs with Timestamp input
select unix_timestamp(t), year(t), month(t), day(t), dayofmonth(t),
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/transform1.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/transform1.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/transform1.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/transform1.q Tue Nov 5 07:01:32 2013
@@ -12,7 +12,7 @@ SELECT transform(*) USING 'cat' AS (col
create table transform1_t2(col array<int>);
insert overwrite table transform1_t2
-select array(1,2,3) from src limit 1;
+select array(1,2,3) from src tablesample (1 rows);
EXPLAIN
SELECT transform('0\0021\0022') USING 'cat' AS (col array<int>) FROM transform1_t2;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/truncate_column.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/truncate_column.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/truncate_column.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/truncate_column.q Tue Nov 5 07:01:32 2013
@@ -5,7 +5,7 @@ ROW FORMAT SERDE 'org.apache.hadoop.hive
set hive.stats.autogather=true;
-INSERT OVERWRITE TABLE test_tab SELECT * FROM src LIMIT 10;
+INSERT OVERWRITE TABLE test_tab SELECT * FROM src tablesample (10 rows);
DESC FORMATTED test_tab;
@@ -20,7 +20,7 @@ DESC FORMATTED test_tab;
SELECT * FROM test_tab ORDER BY value;
-- Truncate multiple columns
-INSERT OVERWRITE TABLE test_tab SELECT * FROM src LIMIT 10;
+INSERT OVERWRITE TABLE test_tab SELECT * FROM src tablesample (10 rows);
TRUNCATE TABLE test_tab COLUMNS (key, value);
@@ -40,7 +40,7 @@ SELECT * FROM test_tab ORDER BY value;
-- Test truncating with a binary serde
ALTER TABLE test_tab SET SERDE 'org.apache.hadoop.hive.serde2.columnar.LazyBinaryColumnarSerDe';
-INSERT OVERWRITE TABLE test_tab SELECT * FROM src LIMIT 10;
+INSERT OVERWRITE TABLE test_tab SELECT * FROM src tablesample (10 rows);
DESC FORMATTED test_tab;
@@ -65,7 +65,7 @@ SELECT * FROM test_tab ORDER BY value;
-- Test truncating a partition
CREATE TABLE test_tab_part (key STRING, value STRING) PARTITIONED BY (part STRING) STORED AS RCFILE;
-INSERT OVERWRITE TABLE test_tab_part PARTITION (part = '1') SELECT * FROM src LIMIT 10;
+INSERT OVERWRITE TABLE test_tab_part PARTITION (part = '1') SELECT * FROM src tablesample (10 rows);
DESC FORMATTED test_tab_part PARTITION (part = '1');
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/truncate_column_merge.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/truncate_column_merge.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/truncate_column_merge.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/truncate_column_merge.q Tue Nov 5 07:01:32 2013
@@ -2,9 +2,9 @@
CREATE TABLE test_tab (key STRING, value STRING) STORED AS RCFILE;
-INSERT OVERWRITE TABLE test_tab SELECT * FROM src LIMIT 5;
+INSERT OVERWRITE TABLE test_tab SELECT * FROM src tablesample (5 rows);
-INSERT INTO TABLE test_tab SELECT * FROM src LIMIT 5;
+INSERT INTO TABLE test_tab SELECT * FROM src tablesample (5 rows);
-- The value should be 2 indicating the table has 2 files
SELECT COUNT(DISTINCT INPUT__FILE__NAME) FROM test_tab;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/truncate_table.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/truncate_table.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/truncate_table.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/truncate_table.q Tue Nov 5 07:01:32 2013
@@ -1,5 +1,5 @@
create table src_truncate (key string, value string);
-load data local inpath '../data/files/kv1.txt' into table src_truncate;;
+load data local inpath '../../data/files/kv1.txt' into table src_truncate;;
create table srcpart_truncate (key string, value string) partitioned by (ds string, hr string);
alter table srcpart_truncate add partition (ds='2008-04-08', hr='11');
@@ -7,10 +7,10 @@ alter table srcpart_truncate add partiti
alter table srcpart_truncate add partition (ds='2008-04-09', hr='11');
alter table srcpart_truncate add partition (ds='2008-04-09', hr='12');
-load data local inpath '../data/files/kv1.txt' into table srcpart_truncate partition (ds='2008-04-08', hr='11');
-load data local inpath '../data/files/kv1.txt' into table srcpart_truncate partition (ds='2008-04-08', hr='12');
-load data local inpath '../data/files/kv1.txt' into table srcpart_truncate partition (ds='2008-04-09', hr='11');
-load data local inpath '../data/files/kv1.txt' into table srcpart_truncate partition (ds='2008-04-09', hr='12');
+load data local inpath '../../data/files/kv1.txt' into table srcpart_truncate partition (ds='2008-04-08', hr='11');
+load data local inpath '../../data/files/kv1.txt' into table srcpart_truncate partition (ds='2008-04-08', hr='12');
+load data local inpath '../../data/files/kv1.txt' into table srcpart_truncate partition (ds='2008-04-09', hr='11');
+load data local inpath '../../data/files/kv1.txt' into table srcpart_truncate partition (ds='2008-04-09', hr='12');
set hive.fetch.task.convertion=more;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/type_cast_1.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/type_cast_1.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/type_cast_1.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/type_cast_1.q Tue Nov 5 07:01:32 2013
@@ -1,3 +1,5 @@
+set hive.fetch.task.conversion=more;
+
EXPLAIN
SELECT IF(false, 1, cast(2 as smallint)) + 3 FROM src LIMIT 1;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/type_conversions_1.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/type_conversions_1.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/type_conversions_1.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/type_conversions_1.q Tue Nov 5 07:01:32 2013
@@ -1,3 +1,4 @@
+set hive.fetch.task.conversion=more;
-- casting from null should yield null
select
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/type_widening.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/type_widening.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/type_widening.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/type_widening.q Tue Nov 5 07:01:32 2013
@@ -1,3 +1,4 @@
+set hive.fetch.task.conversion=more;
-- Check for int, bigint automatic type widening conversions in UDFs, UNIONS
EXPLAIN SELECT COALESCE(0, 9223372036854775807) FROM src LIMIT 1;
SELECT COALESCE(0, 9223372036854775807) FROM src LIMIT 1;
Modified: hive/branches/tez/ql/src/test/queries/clientpositive/udaf_context_ngrams.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/udaf_context_ngrams.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/udaf_context_ngrams.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/udaf_context_ngrams.q Tue Nov 5 07:01:32 2013
@@ -1,5 +1,5 @@
CREATE TABLE kafka (contents STRING);
-LOAD DATA LOCAL INPATH '../data/files/text-en.txt' INTO TABLE kafka;
+LOAD DATA LOCAL INPATH '../../data/files/text-en.txt' INTO TABLE kafka;
set mapred.reduce.tasks=1;
set hive.exec.reducers.max=1;