You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by gu...@apache.org on 2013/11/05 08:01:58 UTC

svn commit: r1538880 [12/46] - in /hive/branches/tez: ./ ant/ ant/src/org/apache/hadoop/hive/ant/ beeline/ beeline/src/java/org/apache/hive/beeline/ beeline/src/main/ beeline/src/test/org/apache/hive/beeline/src/test/ cli/ common/ common/src/java/conf/...

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/progress_1.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/progress_1.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/progress_1.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/progress_1.q Tue Nov  5 07:01:32 2013
@@ -2,7 +2,7 @@ set hive.heartbeat.interval=5; 
 
 
 CREATE TABLE PROGRESS_1(key int, value string) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/kv6.txt' INTO TABLE PROGRESS_1;
+LOAD DATA LOCAL INPATH '../../data/files/kv6.txt' INTO TABLE PROGRESS_1;
 
 select count(1) from PROGRESS_1 t1 join PROGRESS_1 t2 on t1.key=t2.key;
 

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/ptf.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/ptf.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/ptf.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/ptf.q Tue Nov  5 07:01:32 2013
@@ -13,7 +13,7 @@ CREATE TABLE part( 
     p_comment STRING
 );
 
-LOAD DATA LOCAL INPATH '../data/files/part_tiny.txt' overwrite into table part;
+LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part;
 
 --1. test1
 select p_mfgr, p_name, p_size,

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/ptf_decimal.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/ptf_decimal.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/ptf_decimal.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/ptf_decimal.q Tue Nov  5 07:01:32 2013
@@ -13,7 +13,7 @@ CREATE TABLE part( 
     p_comment STRING
 );
 
-LOAD DATA LOCAL INPATH '../data/files/part_tiny.txt' overwrite into table part;
+LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part;
 
 -- 1. aggregate functions with decimal type
 

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/ptf_general_queries.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/ptf_general_queries.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/ptf_general_queries.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/ptf_general_queries.q Tue Nov  5 07:01:32 2013
@@ -13,7 +13,7 @@ CREATE TABLE part( 
     p_comment STRING
 );
 
-LOAD DATA LOCAL INPATH '../data/files/part_tiny.txt' overwrite into table part;
+LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part;
 
 -- 1. testNoPTFNoWindowing
 select p_mfgr, p_name, p_size

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/ptf_matchpath.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/ptf_matchpath.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/ptf_matchpath.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/ptf_matchpath.q Tue Nov  5 07:01:32 2013
@@ -10,7 +10,7 @@ ARR_DELAY float, 
 FL_NUM string 
 );
 
-LOAD DATA LOCAL INPATH '../data/files/flights_tiny.txt' OVERWRITE INTO TABLE flights_tiny;
+LOAD DATA LOCAL INPATH '../../data/files/flights_tiny.txt' OVERWRITE INTO TABLE flights_tiny;
 
 -- 1. basic Matchpath test
 select origin_city_name, fl_num, year, month, day_of_month, sz, tpath 

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/ptf_rcfile.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/ptf_rcfile.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/ptf_rcfile.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/ptf_rcfile.q Tue Nov  5 07:01:32 2013
@@ -12,7 +12,7 @@ CREATE TABLE part_rc( 
     p_comment STRING
 )  STORED AS RCFILE ;
 
-LOAD DATA LOCAL INPATH '../data/files/part.rc' overwrite into table part_rc;
+LOAD DATA LOCAL INPATH '../../data/files/part.rc' overwrite into table part_rc;
 
 -- testWindowingPTFWithPartRC
 select p_mfgr, p_name, p_size, 

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/ptf_register_tblfn.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/ptf_register_tblfn.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/ptf_register_tblfn.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/ptf_register_tblfn.q Tue Nov  5 07:01:32 2013
@@ -10,7 +10,7 @@ ARR_DELAY float, 
 FL_NUM string 
 );
 
-LOAD DATA LOCAL INPATH '../data/files/flights_tiny.txt' OVERWRITE INTO TABLE flights_tiny;
+LOAD DATA LOCAL INPATH '../../data/files/flights_tiny.txt' OVERWRITE INTO TABLE flights_tiny;
 
 create temporary function matchpathtest as 'org.apache.hadoop.hive.ql.udf.ptf.MatchPath$MatchPathResolver';
 

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/ptf_seqfile.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/ptf_seqfile.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/ptf_seqfile.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/ptf_seqfile.q Tue Nov  5 07:01:32 2013
@@ -12,7 +12,7 @@ CREATE TABLE part_seq( 
     p_comment STRING
 ) STORED AS SEQUENCEFILE ;
 
-LOAD DATA LOCAL INPATH '../data/files/part.seq' overwrite into table part_seq;
+LOAD DATA LOCAL INPATH '../../data/files/part.seq' overwrite into table part_seq;
 
 -- testWindowingPTFWithPartSeqFile
 select p_mfgr, p_name, p_size, 

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/ql_rewrite_gbtoidx.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/ql_rewrite_gbtoidx.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/ql_rewrite_gbtoidx.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/ql_rewrite_gbtoidx.q Tue Nov  5 07:01:32 2013
@@ -19,7 +19,7 @@ CREATE TABLE lineitem (L_ORDERKEY      I
 ROW FORMAT DELIMITED
 FIELDS TERMINATED BY '|';
 
-LOAD DATA LOCAL INPATH '../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem;
+LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem;
 
 CREATE INDEX lineitem_lshipdate_idx ON TABLE lineitem(l_shipdate) AS 'org.apache.hadoop.hive.ql.index.AggregateIndexHandler' WITH DEFERRED REBUILD IDXPROPERTIES("AGGREGATES"="count(l_shipdate)");
 ALTER INDEX lineitem_lshipdate_idx ON lineitem REBUILD;
@@ -156,7 +156,7 @@ DROP INDEX tbl_part_index on tblpart;
 DROP TABLE tblpart;
 
 CREATE TABLE tbl(key int, value int) ROW FORMAT DELIMITED FIELDS TERMINATED BY '|'; 
-LOAD DATA LOCAL INPATH '../data/files/tbl.txt' OVERWRITE INTO TABLE tbl;
+LOAD DATA LOCAL INPATH '../../data/files/tbl.txt' OVERWRITE INTO TABLE tbl;
 
 CREATE INDEX tbl_key_idx ON TABLE tbl(key) AS 'org.apache.hadoop.hive.ql.index.AggregateIndexHandler' WITH DEFERRED REBUILD IDXPROPERTIES("AGGREGATES"="count(key)");
 ALTER INDEX tbl_key_idx ON tbl REBUILD;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/quote2.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/quote2.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/quote2.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/quote2.q Tue Nov  5 07:01:32 2013
@@ -1,3 +1,5 @@
+set hive.fetch.task.conversion=more;
+
 EXPLAIN
 SELECT
     'abc',        "abc",

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/rcfile_bigdata.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/rcfile_bigdata.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/rcfile_bigdata.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/rcfile_bigdata.q Tue Nov  5 07:01:32 2013
@@ -1,7 +1,7 @@
 set hive.map.aggr.hash.percentmemory = 0.3;
 set hive.mapred.local.mem = 256;
 
-add file ../data/scripts/dumpdata_script.py;
+add file ../../data/scripts/dumpdata_script.py;
 
 CREATE table columnTable_Bigdata (key STRING, value STRING)
 ROW FORMAT SERDE

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/remote_script.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/remote_script.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/remote_script.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/remote_script.q Tue Nov  5 07:01:32 2013
@@ -1,4 +1,4 @@
-dfs -put ../data/scripts/newline.py /newline.py;
+dfs -put ../../data/scripts/newline.py /newline.py;
 add file hdfs:///newline.py;
 set hive.transform.escape.input=true;
 

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/repair.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/repair.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/repair.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/repair.q Tue Nov  5 07:01:32 2013
@@ -1,10 +1,12 @@
+DROP TABLE IF EXISTS repairtable;
+
 CREATE TABLE repairtable(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
 
 MSCK TABLE repairtable;
 
-dfs ${system:test.dfs.mkdir} ../build/ql/test/data/warehouse/repairtable/p1=a/p2=a;
-dfs ${system:test.dfs.mkdir} ../build/ql/test/data/warehouse/repairtable/p1=b/p2=a;
-dfs -touchz ../build/ql/test/data/warehouse/repairtable/p1=b/p2=a/datafile;
+dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=a/p2=a;
+dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=b/p2=a;
+dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=b/p2=a/datafile;
 
 MSCK TABLE repairtable;
 
@@ -12,4 +14,4 @@ MSCK REPAIR TABLE repairtable;
 
 MSCK TABLE repairtable;
 
-
+DROP TABLE repairtable;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/scriptfile1.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/scriptfile1.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/scriptfile1.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/scriptfile1.q Tue Nov  5 07:01:32 2013
@@ -3,7 +3,7 @@ set hive.input.format=org.apache.hadoop.
 -- EXCLUDE_OS_WINDOWS
 CREATE TABLE dest1(key INT, value STRING);
 
-ADD FILE src/test/scripts/testgrep;
+ADD FILE ../../ql/src/test/scripts/testgrep;
 
 FROM (
   FROM src

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/serde_regex.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/serde_regex.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/serde_regex.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/serde_regex.q Tue Nov  5 07:01:32 2013
@@ -31,8 +31,8 @@ WITH SERDEPROPERTIES (
 )
 STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH "../data/files/apache.access.log" INTO TABLE serde_regex;
-LOAD DATA LOCAL INPATH "../data/files/apache.access.2.log" INTO TABLE serde_regex;
+LOAD DATA LOCAL INPATH "../../data/files/apache.access.log" INTO TABLE serde_regex;
+LOAD DATA LOCAL INPATH "../../data/files/apache.access.2.log" INTO TABLE serde_regex;
 
 SELECT * FROM serde_regex ORDER BY time;
 
@@ -59,7 +59,7 @@ WITH SERDEPROPERTIES (
 )
 STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH "../data/files/kv7.txt" INTO TABLE serde_regex1;
+LOAD DATA LOCAL INPATH "../../data/files/kv7.txt" INTO TABLE serde_regex1;
 
 SELECT key, value FROM serde_regex1 ORDER BY key, value;
 

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/set_processor_namespaces.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/set_processor_namespaces.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/set_processor_namespaces.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/set_processor_namespaces.q Tue Nov  5 07:01:32 2013
@@ -24,7 +24,7 @@ set b=a;
 set c=${hiveconf:${hiveconf:b}};
 set c;
 
-set jar=${system:build.ivy.lib.dir}/default/derby-${system:derby.version}.jar;
+set jar=${system:maven.local.repository}/org/apache/derby/derby/${system:derby.version}/derby-${system:derby.version}.jar;
 
 add file ${hiveconf:jar};
 delete file ${hiveconf:jar};

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin.q Tue Nov  5 07:01:32 2013
@@ -13,10 +13,10 @@ CREATE TABLE T3(key STRING, val STRING) 
 CREATE TABLE T4(key STRING, val STRING) STORED AS TEXTFILE;
 CREATE TABLE dest_j1(key INT, value STRING) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
-LOAD DATA LOCAL INPATH '../data/files/T3.txt' INTO TABLE T3;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T4;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T4;
 
 
 EXPLAIN

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin_union_remove_1.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin_union_remove_1.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin_union_remove_1.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin_union_remove_1.q Tue Nov  5 07:01:32 2013
@@ -20,12 +20,12 @@ set mapred.input.dir.recursive=true;
 CREATE TABLE T1(key STRING, val STRING)
 SKEWED BY (key) ON ((2)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
 
 CREATE TABLE T2(key STRING, val STRING)
 SKEWED BY (key) ON ((3)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
 
 -- a simple join query with skew on both the tables on the join key
 

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin_union_remove_2.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin_union_remove_2.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin_union_remove_2.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoin_union_remove_2.q Tue Nov  5 07:01:32 2013
@@ -12,16 +12,16 @@ set mapred.input.dir.recursive=true;
 CREATE TABLE T1(key STRING, val STRING)
 SKEWED BY (key) ON ((2), (8)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
 
 CREATE TABLE T2(key STRING, val STRING)
 SKEWED BY (key) ON ((3), (8)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
 
 CREATE TABLE T3(key STRING, val STRING) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T3.txt' INTO TABLE T3;
+LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3;
 
 -- This is to test the union->selectstar->filesink and skewjoin optimization
 -- Union of 3 map-reduce subqueries is performed for the skew join

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt1.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt1.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt1.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt1.q Tue Nov  5 07:01:32 2013
@@ -5,12 +5,12 @@ set hive.optimize.skewjoin.compiletime =
 CREATE TABLE T1(key STRING, val STRING)
 SKEWED BY (key) ON ((2)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
 
 CREATE TABLE T2(key STRING, val STRING)
 SKEWED BY (key) ON ((3)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
 
 -- a simple join query with skew on both the tables on the join key
 -- adding a order by at the end to make the results deterministic

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt10.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt10.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt10.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt10.q Tue Nov  5 07:01:32 2013
@@ -4,7 +4,7 @@ set hive.optimize.skewjoin.compiletime =
 
 CREATE TABLE T1(key STRING, value STRING) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
 
 drop table array_valued_T1;
 create table array_valued_T1 (key string, value array<string>) SKEWED BY (key) ON ((8));

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt11.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt11.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt11.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt11.q Tue Nov  5 07:01:32 2013
@@ -5,11 +5,11 @@ set hive.optimize.skewjoin.compiletime =
 CREATE TABLE T1(key STRING, val STRING)	
 SKEWED BY (key) ON ((2)) STORED AS TEXTFILE;	
        
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;	
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;	
      
 CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;	
        
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;	
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;	
      
 -- This test is to verify the skew join compile optimization when the join is followed
 -- by a union. Both sides of a union consist of a join, which should have used

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt12.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt12.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt12.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt12.q Tue Nov  5 07:01:32 2013
@@ -5,12 +5,12 @@ set hive.optimize.skewjoin.compiletime =
 CREATE TABLE T1(key STRING, val STRING)
 SKEWED BY (key, val) ON ((2, 12), (8, 18)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
 
 CREATE TABLE T2(key STRING, val STRING)
 SKEWED BY (key, val) ON ((3, 13), (8, 18)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
 
 -- Both the join tables are skewed by 2 keys, and one of the skewed values
 -- is common to both the tables. The join key matches the skewed key set.

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt13.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt13.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt13.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt13.q Tue Nov  5 07:01:32 2013
@@ -4,16 +4,16 @@ set hive.optimize.skewjoin.compiletime =
 
 CREATE TABLE T1(key STRING, val STRING) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
 
 CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
 
 CREATE TABLE T3(key STRING, val STRING)
 SKEWED BY (val) ON ((12)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T3.txt' INTO TABLE T3;
+LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3;
 
 -- This test is for skewed join compile time optimization for more than 2 tables.
 -- The join key for table 3 is different from the join key used for joining

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt14.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt14.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt14.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt14.q Tue Nov  5 07:01:32 2013
@@ -5,16 +5,16 @@ set hive.optimize.skewjoin.compiletime =
 CREATE TABLE T1(key STRING, val STRING)
 SKEWED BY (key) ON ((2)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
 
 CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
 
 CREATE TABLE T3(key STRING, val STRING)
 SKEWED BY (val) ON ((12)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T3.txt' INTO TABLE T3;
+LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3;
 
 -- This test is for skewed join compile time optimization for more than 2 tables.
 -- The join key for table 3 is different from the join key used for joining

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt15.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt15.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt15.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt15.q Tue Nov  5 07:01:32 2013
@@ -4,7 +4,7 @@ set hive.optimize.skewjoin.compiletime =
 
 CREATE TABLE tmpT1(key STRING, val STRING) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE tmpT1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE tmpT1;
 
 -- testing skew on other data types - int
 CREATE TABLE T1(key INT, val STRING) SKEWED BY (key) ON ((2));
@@ -12,7 +12,7 @@ INSERT OVERWRITE TABLE T1 SELECT key, va
 
 CREATE TABLE tmpT2(key STRING, val STRING) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE tmpT2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE tmpT2;
 
 CREATE TABLE T2(key INT, val STRING) SKEWED BY (key) ON ((3));
 

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt16.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt16.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt16.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt16.q Tue Nov  5 07:01:32 2013
@@ -5,12 +5,12 @@ set hive.optimize.skewjoin.compiletime =
 CREATE TABLE T1(key STRING, val STRING)
 SKEWED BY (key, val) ON ((2, 12)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
 
 CREATE TABLE T2(key STRING, val STRING)
 SKEWED BY (key) ON ((3)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
 
 -- One of the tables is skewed by 2 columns, and the other table is
 -- skewed by one column. Ths join is performed on the both the columns

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt17.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt17.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt17.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt17.q Tue Nov  5 07:01:32 2013
@@ -5,12 +5,12 @@ set hive.optimize.skewjoin.compiletime =
 CREATE TABLE T1(key STRING, val STRING)
 SKEWED BY (key, val) ON ((2, 12)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
 
 CREATE TABLE T2(key STRING, val STRING)
 SKEWED BY (key) ON ((2)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
 
 -- One of the tables is skewed by 2 columns, and the other table is
 -- skewed by one column. Ths join is performed on the first skewed column
@@ -31,12 +31,12 @@ DROP TABLE T2;
 CREATE TABLE T1(key STRING, val STRING)
 SKEWED BY (key, val) ON ((2, 12)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
 
 CREATE TABLE T2(key STRING, val STRING)
 SKEWED BY (key) ON ((2)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
 
 -- One of the tables is skewed by 2 columns, and the other table is
 -- skewed by one column. Ths join is performed on the both the columns

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt18.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt18.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt18.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt18.q Tue Nov  5 07:01:32 2013
@@ -4,7 +4,7 @@ set hive.optimize.skewjoin.compiletime =
 
 CREATE TABLE tmpT1(key STRING, val STRING) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE tmpT1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE tmpT1;
 
 -- testing skew on other data types - int
 CREATE TABLE T1(key INT, val STRING) SKEWED BY (key) ON ((2));
@@ -16,7 +16,7 @@ INSERT OVERWRITE TABLE T1 SELECT key, va
 CREATE TABLE T2(key STRING, val STRING)
 SKEWED BY (key) ON ((3)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
 
 -- Once HIVE-3445 is fixed, the compile time skew join optimization would be
 -- applicable here. Till the above jira is fixed, it would be performed as a

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt19.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt19.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt19.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt19.q Tue Nov  5 07:01:32 2013
@@ -6,11 +6,11 @@ CREATE TABLE T1(key STRING, val STRING)
 CLUSTERED BY (key) INTO 4 BUCKETS
 SKEWED BY (key) ON ((2)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
 
 CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
 
 -- add a test where the skewed key is also the bucketized key
 -- it should not matter, and the compile time skewed join

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt2.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt2.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt2.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt2.q Tue Nov  5 07:01:32 2013
@@ -5,12 +5,12 @@ set hive.optimize.skewjoin.compiletime =
 CREATE TABLE T1(key STRING, val STRING)
 SKEWED BY (key) ON ((2), (7)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
 
 CREATE TABLE T2(key STRING, val STRING)
 SKEWED BY (key) ON ((3), (8)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
 
 -- a simple query with skew on both the tables on the join key
 -- multiple skew values are present for the skewed keys

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt20.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt20.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt20.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt20.q Tue Nov  5 07:01:32 2013
@@ -6,11 +6,11 @@ CREATE TABLE T1(key STRING, val STRING)
 CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS
 SKEWED BY (key) ON ((2)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
 
 CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
 
 -- add a test where the skewed key is also the bucketized/sorted key
 -- it should not matter, and the compile time skewed join

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt3.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt3.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt3.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt3.q Tue Nov  5 07:01:32 2013
@@ -5,12 +5,12 @@ set hive.optimize.skewjoin.compiletime =
 CREATE TABLE T1(key STRING, val STRING)
 SKEWED BY (key) ON ((2), (8)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
 
 CREATE TABLE T2(key STRING, val STRING)
 SKEWED BY (key) ON ((3), (8)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
 
 -- a simple query with skew on both the tables. One of the skewed
 -- value is common to both the tables. The skewed value should not be

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt4.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt4.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt4.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt4.q Tue Nov  5 07:01:32 2013
@@ -5,11 +5,11 @@ set hive.optimize.skewjoin.compiletime =
 CREATE TABLE T1(key STRING, val STRING)
 SKEWED BY (key) ON ((2)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
 
 CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
 
 -- only of the tables of the join (the left table of the join) is skewed
 -- the skewed filter would still be applied to both the tables

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt5.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt5.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt5.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt5.q Tue Nov  5 07:01:32 2013
@@ -5,12 +5,12 @@ set hive.optimize.skewjoin.compiletime =
 CREATE TABLE T1(key STRING, val STRING)
 SKEWED BY (key, val) ON ((2, 12)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
 
 CREATE TABLE T2(key STRING, val STRING)
 SKEWED BY (key) ON ((3)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
 
 -- One of the tables is skewed by 2 columns, and the other table is
 -- skewed by one column. Ths join is performed on the first skewed column

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt6.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt6.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt6.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt6.q Tue Nov  5 07:01:32 2013
@@ -5,12 +5,12 @@ set hive.optimize.skewjoin.compiletime =
 CREATE TABLE T1(key STRING, val STRING)
 SKEWED BY (key, val) ON ((2, 12), (8, 18)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
 
 CREATE TABLE T2(key STRING, val STRING)
 SKEWED BY (key, val) ON ((3, 13), (8, 18)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
 
 -- Both the join tables are skewed by 2 keys, and one of the skewed values
 -- is common to both the tables. The join key is a subset of the skewed key set:

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt7.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt7.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt7.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt7.q Tue Nov  5 07:01:32 2013
@@ -5,16 +5,16 @@ set hive.optimize.skewjoin.compiletime =
 CREATE TABLE T1(key STRING, val STRING)
 SKEWED BY (key) ON ((2), (8)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
 
 CREATE TABLE T2(key STRING, val STRING)
 SKEWED BY (key) ON ((3), (8)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
 
 CREATE TABLE T3(key STRING, val STRING) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T3.txt' INTO TABLE T3;
+LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3;
 
 -- This test is for validating skewed join compile time optimization for more than
 -- 2 tables. The join key is the same, and so a 3-way join would be performed.

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt8.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt8.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt8.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt8.q Tue Nov  5 07:01:32 2013
@@ -4,16 +4,16 @@ set hive.optimize.skewjoin.compiletime =
 
 CREATE TABLE T1(key STRING, val STRING) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
 
 CREATE TABLE T2(key STRING, val STRING)
 SKEWED BY (key) ON ((3), (8)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
 
 CREATE TABLE T3(key STRING, val STRING) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T3.txt' INTO TABLE T3;
+LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3;
 
 -- This test is for validating skewed join compile time optimization for more than
 -- 2 tables. The join key is the same, and so a 3-way join would be performed.

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt9.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt9.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt9.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/skewjoinopt9.q Tue Nov  5 07:01:32 2013
@@ -5,11 +5,11 @@ set hive.optimize.skewjoin.compiletime =
 CREATE TABLE T1(key STRING, val STRING)
 SKEWED BY (key) ON ((2)) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
 
 CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/T2.txt' INTO TABLE T2;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
 
 -- no skew join compile time optimization would be performed if one of the
 -- join sources is a sub-query consisting of a union all

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_1.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_1.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_1.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_1.q Tue Nov  5 07:01:32 2013
@@ -6,9 +6,9 @@ create table smb_bucket_1(key int, value
 create table smb_bucket_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE; 
 create table smb_bucket_3(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
 
-load data local inpath '../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
-load data local inpath '../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
-load data local inpath '../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
+load data local inpath '../../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
+load data local inpath '../../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
+load data local inpath '../../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
 
 set hive.optimize.bucketmapjoin = true;
 set hive.optimize.bucketmapjoin.sortedmerge = true;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_10.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_10.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_10.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_10.q Tue Nov  5 07:01:32 2013
@@ -6,11 +6,11 @@ alter table tmp_smb_bucket_10 add partit
 
 -- add dummy files to make sure that the number of files in each partition is same as number of buckets
  
-load data local inpath '../data/files/smbbucket_1.rc' INTO TABLE tmp_smb_bucket_10 partition(ds='1');
-load data local inpath '../data/files/smbbucket_2.rc' INTO TABLE tmp_smb_bucket_10 partition(ds='1');
+load data local inpath '../../data/files/smbbucket_1.rc' INTO TABLE tmp_smb_bucket_10 partition(ds='1');
+load data local inpath '../../data/files/smbbucket_2.rc' INTO TABLE tmp_smb_bucket_10 partition(ds='1');
 
-load data local inpath '../data/files/smbbucket_1.rc' INTO TABLE tmp_smb_bucket_10 partition(ds='2');
-load data local inpath '../data/files/smbbucket_2.rc' INTO TABLE tmp_smb_bucket_10 partition(ds='2');
+load data local inpath '../../data/files/smbbucket_1.rc' INTO TABLE tmp_smb_bucket_10 partition(ds='2');
+load data local inpath '../../data/files/smbbucket_2.rc' INTO TABLE tmp_smb_bucket_10 partition(ds='2');
 
 set hive.optimize.bucketmapjoin = true;
 set hive.optimize.bucketmapjoin.sortedmerge = true;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_2.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_2.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_2.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_2.q Tue Nov  5 07:01:32 2013
@@ -6,9 +6,9 @@ create table smb_bucket_1(key int, value
 create table smb_bucket_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE; 
 create table smb_bucket_3(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
 
-load data local inpath '../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
-load data local inpath '../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
-load data local inpath '../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
+load data local inpath '../../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
+load data local inpath '../../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
+load data local inpath '../../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
  
 set hive.optimize.bucketmapjoin = true;
 set hive.optimize.bucketmapjoin.sortedmerge = true;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_25.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_25.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_25.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_25.q Tue Nov  5 07:01:32 2013
@@ -10,9 +10,9 @@ create table smb_bucket_1(key int, value
 create table smb_bucket_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE; 
 create table smb_bucket_3(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
 
-load data local inpath '../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
-load data local inpath '../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
-load data local inpath '../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
+load data local inpath '../../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
+load data local inpath '../../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
+load data local inpath '../../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
 
 explain 
 select * from (select a.key from smb_bucket_1 a join smb_bucket_2 b on (a.key = b.key) where a.key = 5) t1 left outer join (select c.key from smb_bucket_2 c join smb_bucket_3 d on (c.key = d.key) where c.key=5) t2 on (t1.key=t2.key) where t2.key=5;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_3.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_3.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_3.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_3.q Tue Nov  5 07:01:32 2013
@@ -6,9 +6,9 @@ create table smb_bucket_1(key int, value
 create table smb_bucket_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE; 
 create table smb_bucket_3(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
 
-load data local inpath '../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
-load data local inpath '../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
-load data local inpath '../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
+load data local inpath '../../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
+load data local inpath '../../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
+load data local inpath '../../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
 
 set hive.optimize.bucketmapjoin = true;
 set hive.optimize.bucketmapjoin.sortedmerge = true;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_4.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_4.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_4.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_4.q Tue Nov  5 07:01:32 2013
@@ -6,9 +6,9 @@ create table smb_bucket_1(key int, value
 create table smb_bucket_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE; 
 create table smb_bucket_3(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
 
-load data local inpath '../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
-load data local inpath '../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
-load data local inpath '../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
+load data local inpath '../../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
+load data local inpath '../../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
+load data local inpath '../../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
 
 set hive.optimize.bucketmapjoin = true;
 set hive.optimize.bucketmapjoin.sortedmerge = true;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_5.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_5.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_5.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_5.q Tue Nov  5 07:01:32 2013
@@ -6,9 +6,9 @@ create table smb_bucket_1(key int, value
 create table smb_bucket_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE; 
 create table smb_bucket_3(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
 
-load data local inpath '../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
-load data local inpath '../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
-load data local inpath '../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
+load data local inpath '../../data/files/smbbucket_1.rc' overwrite into table smb_bucket_1;
+load data local inpath '../../data/files/smbbucket_2.rc' overwrite into table smb_bucket_2;
+load data local inpath '../../data/files/smbbucket_3.rc' overwrite into table smb_bucket_3;
 
 set hive.optimize.bucketmapjoin = true;
 set hive.optimize.bucketmapjoin.sortedmerge = true;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_7.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_7.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_7.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_7.q Tue Nov  5 07:01:32 2013
@@ -15,8 +15,8 @@ create table smb_join_results(k1 int, v1
 create table smb_join_results_empty_bigtable(k1 int, v1 string, k2 int, v2 string);
 create table normal_join_results(k1 int, v1 string, k2 int, v2 string);
 
-load data local inpath '../data/files/empty1.txt' into table smb_bucket4_1;
-load data local inpath '../data/files/empty2.txt' into table smb_bucket4_1;
+load data local inpath '../../data/files/empty1.txt' into table smb_bucket4_1;
+load data local inpath '../../data/files/empty2.txt' into table smb_bucket4_1;
 
 insert overwrite table smb_bucket4_2
 select * from src;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_8.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_8.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_8.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/smb_mapjoin_8.q Tue Nov  5 07:01:32 2013
@@ -5,7 +5,7 @@ set hive.exec.reducers.max = 1;
 
 
 create table smb_bucket_input (key int, value string) stored as rcfile;
-load data local inpath '../data/files/smb_bucket_input.rc' into table smb_bucket_input;
+load data local inpath '../../data/files/smb_bucket_input.rc' into table smb_bucket_input;
 
 
 CREATE TABLE smb_bucket4_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/source.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/source.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/source.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/source.q Tue Nov  5 07:01:32 2013
@@ -1 +1 @@
-source ../data/files/source.txt;
+source ../../data/files/source.txt;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/split.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/split.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/split.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/split.q Tue Nov  5 07:01:32 2013
@@ -2,7 +2,7 @@ DROP TABLE tmp_jo_tab_test;
 CREATE table tmp_jo_tab_test (message_line STRING)
 STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../data/files/input.txt'
+LOAD DATA LOCAL INPATH '../../data/files/input.txt'
 OVERWRITE INTO TABLE tmp_jo_tab_test;
 
 select size(split(message_line, '\t')) from tmp_jo_tab_test;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/stats1.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/stats1.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/stats1.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/stats1.q Tue Nov  5 07:01:32 2013
@@ -26,5 +26,5 @@ DESCRIBE FORMATTED tmptable;
 -- Load a file into a existing table
 -- Some stats (numFiles, totalSize) should be updated correctly
 -- Some other stats (numRows, rawDataSize) should be cleared
-load data local inpath '../data/files/srcbucket20.txt' INTO TABLE tmptable;
+load data local inpath '../../data/files/srcbucket20.txt' INTO TABLE tmptable;
 DESCRIBE FORMATTED tmptable;
\ No newline at end of file

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/stats11.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/stats11.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/stats11.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/stats11.q Tue Nov  5 07:01:32 2013
@@ -2,25 +2,25 @@ set datanucleus.cache.collections=false;
 set hive.stats.autogather=true;
 
 CREATE TABLE srcbucket_mapjoin(key int, value string) CLUSTERED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE;
-load data local inpath '../data/files/srcbucket20.txt' INTO TABLE srcbucket_mapjoin;
-load data local inpath '../data/files/srcbucket21.txt' INTO TABLE srcbucket_mapjoin;
+load data local inpath '../../data/files/srcbucket20.txt' INTO TABLE srcbucket_mapjoin;
+load data local inpath '../../data/files/srcbucket21.txt' INTO TABLE srcbucket_mapjoin;
 
 CREATE TABLE srcbucket_mapjoin_part (key int, value string) partitioned by (ds string) CLUSTERED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE;
 explain
-load data local inpath '../data/files/srcbucket20.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
-load data local inpath '../data/files/srcbucket20.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
+load data local inpath '../../data/files/srcbucket20.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
+load data local inpath '../../data/files/srcbucket20.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
 
 desc formatted srcbucket_mapjoin_part partition(ds='2008-04-08');
-load data local inpath '../data/files/srcbucket21.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
+load data local inpath '../../data/files/srcbucket21.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
 desc formatted srcbucket_mapjoin_part partition(ds='2008-04-08');
-load data local inpath '../data/files/srcbucket22.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
+load data local inpath '../../data/files/srcbucket22.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
 desc formatted srcbucket_mapjoin_part partition(ds='2008-04-08');
-load data local inpath '../data/files/srcbucket23.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
+load data local inpath '../../data/files/srcbucket23.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08');
 desc formatted srcbucket_mapjoin_part partition(ds='2008-04-08');
 
 CREATE TABLE srcbucket_mapjoin_part_2 (key int, value string) partitioned by (ds string) CLUSTERED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE;
-load data local inpath '../data/files/srcbucket22.txt' INTO TABLE srcbucket_mapjoin_part_2 partition(ds='2008-04-08');
-load data local inpath '../data/files/srcbucket23.txt' INTO TABLE srcbucket_mapjoin_part_2 partition(ds='2008-04-08');
+load data local inpath '../../data/files/srcbucket22.txt' INTO TABLE srcbucket_mapjoin_part_2 partition(ds='2008-04-08');
+load data local inpath '../../data/files/srcbucket23.txt' INTO TABLE srcbucket_mapjoin_part_2 partition(ds='2008-04-08');
 
 create table bucketmapjoin_hash_result_1 (key bigint , value1 bigint, value2 bigint);
 create table bucketmapjoin_hash_result_2 (key bigint , value1 bigint, value2 bigint);

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/stats18.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/stats18.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/stats18.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/stats18.q Tue Nov  5 07:01:32 2013
@@ -13,7 +13,7 @@ insert overwrite table stats_part partit
 -- Some other stats (numRows, rawDataSize) should be cleared
 desc formatted stats_part partition (ds='2010-04-08', hr='13');
 
-load data local inpath '../data/files/srcbucket20.txt' INTO TABLE stats_part partition (ds='2010-04-08', hr='13');
+load data local inpath '../../data/files/srcbucket20.txt' INTO TABLE stats_part partition (ds='2010-04-08', hr='13');
 
 desc formatted stats_part partition (ds='2010-04-08', hr='13');
 

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/stats3.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/stats3.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/stats3.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/stats3.q Tue Nov  5 07:01:32 2013
@@ -5,9 +5,9 @@ drop table hive_test_dst;
 
 create table hive_test_src ( col1 string ) stored as textfile ;
 explain extended
-load data local inpath '../data/files/test.dat' overwrite into table hive_test_src ;
+load data local inpath '../../data/files/test.dat' overwrite into table hive_test_src ;
 
-load data local inpath '../data/files/test.dat' overwrite into table hive_test_src ;
+load data local inpath '../../data/files/test.dat' overwrite into table hive_test_src ;
 
 desc formatted hive_test_src;
 

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/stats_noscan_2.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/stats_noscan_2.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/stats_noscan_2.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/stats_noscan_2.q Tue Nov  5 07:01:32 2013
@@ -1,7 +1,7 @@
 
 -- test analyze table compute statistiscs [noscan] on external table 
 -- 1 test table
-CREATE EXTERNAL TABLE anaylyze_external (a INT) LOCATION '${system:test.src.data.dir}/files/ext_test';
+CREATE EXTERNAL TABLE anaylyze_external (a INT) LOCATION '${system:hive.root}/data/files/ext_test';
 SELECT * FROM anaylyze_external;
 analyze table anaylyze_external compute statistics;
 describe formatted anaylyze_external;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/str_to_map.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/str_to_map.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/str_to_map.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/str_to_map.q Tue Nov  5 07:01:32 2013
@@ -1,3 +1,5 @@
+set hive.fetch.task.conversion=more;
+
 desc function str_to_map;
 desc function extended str_to_map;
 
@@ -19,7 +21,7 @@ limit 3;
 
 
 drop table tbl_s2m;
-create table tbl_s2m as select 'ABC=CC_333=444' as t from src limit 3;
+create table tbl_s2m as select 'ABC=CC_333=444' as t from src tablesample (3 rows);
 
 select str_to_map(t,'_','=')['333'] from tbl_s2m;
 

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/subq.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/subq.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/subq.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/subq.q Tue Nov  5 07:01:32 2013
@@ -2,12 +2,12 @@ EXPLAIN
 FROM (
   FROM src select src.* WHERE src.key < 100
 ) unioninput
-INSERT OVERWRITE DIRECTORY '../build/ql/test/data/warehouse/union.out' SELECT unioninput.*;
+INSERT OVERWRITE DIRECTORY 'target/warehouse/union.out' SELECT unioninput.*;
 
 FROM (
   FROM src select src.* WHERE src.key < 100
 ) unioninput
-INSERT OVERWRITE DIRECTORY '../build/ql/test/data/warehouse/union.out' SELECT unioninput.*;
+INSERT OVERWRITE DIRECTORY 'target/warehouse/union.out' SELECT unioninput.*;
 
-dfs -cat ../build/ql/test/data/warehouse/union.out/*;
+dfs -cat ${system:test.warehouse.dir}/union.out/*;
 

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/subquery_in.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/subquery_in.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/subquery_in.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/subquery_in.q Tue Nov  5 07:01:32 2013
@@ -13,7 +13,7 @@ CREATE TABLE part( 
     p_comment STRING
 );
 
-LOAD DATA LOCAL INPATH '../data/files/part_tiny.txt' overwrite into table part;
+LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part;
 
 DROP TABLE lineitem;
 CREATE TABLE lineitem (L_ORDERKEY      INT,
@@ -35,7 +35,7 @@ CREATE TABLE lineitem (L_ORDERKEY      I
 ROW FORMAT DELIMITED
 FIELDS TERMINATED BY '|';
 
-LOAD DATA LOCAL INPATH '../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem;
+LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem;
 
 -- non agg, non corr
 explain

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/subquery_notin.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/subquery_notin.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/subquery_notin.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/subquery_notin.q Tue Nov  5 07:01:32 2013
@@ -13,7 +13,7 @@ CREATE TABLE part( 
     p_comment STRING
 );
 
-LOAD DATA LOCAL INPATH '../data/files/part_tiny.txt' overwrite into table part;
+LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part;
 
 DROP TABLE lineitem;
 CREATE TABLE lineitem (L_ORDERKEY      INT,
@@ -35,7 +35,7 @@ CREATE TABLE lineitem (L_ORDERKEY      I
 ROW FORMAT DELIMITED
 FIELDS TERMINATED BY '|';
 
-LOAD DATA LOCAL INPATH '../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem;
+LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem;
 
 -- non agg, non corr
 explain

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/symlink_text_input_format.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/symlink_text_input_format.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/symlink_text_input_format.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/symlink_text_input_format.q Tue Nov  5 07:01:32 2013
@@ -1,12 +1,12 @@
-
+DROP TABLE IF EXISTS symlink_text_input_format;
 
 EXPLAIN
 CREATE TABLE symlink_text_input_format (key STRING, value STRING) STORED AS INPUTFORMAT 'org.apache.hadoop.hive.ql.io.SymlinkTextInputFormat' OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat';
 
 CREATE TABLE symlink_text_input_format (key STRING, value STRING) STORED AS INPUTFORMAT 'org.apache.hadoop.hive.ql.io.SymlinkTextInputFormat' OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat';
 
-dfs -cp ../data/files/symlink1.txt ../build/ql/test/data/warehouse/symlink_text_input_format/symlink1.txt;
-dfs -cp ../data/files/symlink2.txt ../build/ql/test/data/warehouse/symlink_text_input_format/symlink2.txt;
+dfs -cp ../../data/files/symlink1.txt ${system:test.warehouse.dir}/symlink_text_input_format/symlink1.txt;
+dfs -cp ../../data/files/symlink2.txt ${system:test.warehouse.dir}/symlink_text_input_format/symlink2.txt;
 
 EXPLAIN SELECT * FROM symlink_text_input_format order by key, value;
 
@@ -20,5 +20,4 @@ EXPLAIN SELECT count(1) FROM symlink_tex
 
 SELECT count(1) FROM symlink_text_input_format;
 
-
-
+DROP TABLE symlink_text_input_format;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/table_access_keys_stats.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/table_access_keys_stats.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/table_access_keys_stats.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/table_access_keys_stats.q Tue Nov  5 07:01:32 2013
@@ -4,7 +4,7 @@ SET hive.stats.collect.tablekeys=true;
 -- This test is used for testing the TableAccessAnalyzer
 
 CREATE TABLE T1(key STRING, val STRING) STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
 
 CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;
 CREATE TABLE T3(key STRING, val STRING) STORED AS TEXTFILE;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/test_boolean_whereclause.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/test_boolean_whereclause.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/test_boolean_whereclause.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/test_boolean_whereclause.q Tue Nov  5 07:01:32 2013
@@ -1,5 +1,5 @@
 create table if not exists test_boolean(dummy tinyint);
-insert overwrite table test_boolean  select 1 from src limit 1;
+insert overwrite table test_boolean  select 1 from src tablesample (1 rows);
 
 SELECT 1
 FROM (

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_1.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_1.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_1.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_1.q Tue Nov  5 07:01:32 2013
@@ -1,10 +1,12 @@
+set hive.fetch.task.conversion=more;
+
 drop table timestamp_1;
 
 create table timestamp_1 (t timestamp);
 alter table timestamp_1 set serde 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe';
 
-insert overwrite table timestamp_1 
-  select cast('2011-01-01 01:01:01' as timestamp) from src limit 1;
+insert overwrite table timestamp_1
+  select cast('2011-01-01 01:01:01' as timestamp) from src tablesample (1 rows);
 select cast(t as boolean) from timestamp_1 limit 1;
 select cast(t as tinyint) from timestamp_1 limit 1;
 select cast(t as smallint) from timestamp_1 limit 1;
@@ -15,7 +17,7 @@ select cast(t as double) from timestamp_
 select cast(t as string) from timestamp_1 limit 1;
 
 insert overwrite table timestamp_1
-  select '2011-01-01 01:01:01' from src limit 1;
+  select '2011-01-01 01:01:01' from src tablesample (1 rows);
 select cast(t as boolean) from timestamp_1 limit 1;
 select cast(t as tinyint) from timestamp_1 limit 1;
 select cast(t as smallint) from timestamp_1 limit 1;
@@ -26,7 +28,7 @@ select cast(t as double) from timestamp_
 select cast(t as string) from timestamp_1 limit 1;
 
 insert overwrite table timestamp_1
-  select '2011-01-01 01:01:01.1' from src limit 1;
+  select '2011-01-01 01:01:01.1' from src tablesample (1 rows);
 select cast(t as boolean) from timestamp_1 limit 1;
 select cast(t as tinyint) from timestamp_1 limit 1;
 select cast(t as smallint) from timestamp_1 limit 1;
@@ -37,7 +39,7 @@ select cast(t as double) from timestamp_
 select cast(t as string) from timestamp_1 limit 1;
 
 insert overwrite table timestamp_1
-  select '2011-01-01 01:01:01.0001' from src limit 1;
+  select '2011-01-01 01:01:01.0001' from src tablesample (1 rows);
 select cast(t as boolean) from timestamp_1 limit 1;
 select cast(t as tinyint) from timestamp_1 limit 1;
 select cast(t as smallint) from timestamp_1 limit 1;
@@ -48,7 +50,7 @@ select cast(t as double) from timestamp_
 select cast(t as string) from timestamp_1 limit 1;
 
 insert overwrite table timestamp_1
-  select '2011-01-01 01:01:01.000100000' from src limit 1;
+  select '2011-01-01 01:01:01.000100000' from src tablesample (1 rows);
 select cast(t as boolean) from timestamp_1 limit 1;
 select cast(t as tinyint) from timestamp_1 limit 1;
 select cast(t as smallint) from timestamp_1 limit 1;
@@ -59,7 +61,7 @@ select cast(t as double) from timestamp_
 select cast(t as string) from timestamp_1 limit 1;
 
 insert overwrite table timestamp_1
-  select '2011-01-01 01:01:01.001000011' from src limit 1;
+  select '2011-01-01 01:01:01.001000011' from src tablesample (1 rows);
 select cast(t as boolean) from timestamp_1 limit 1;
 select cast(t as tinyint) from timestamp_1 limit 1;
 select cast(t as smallint) from timestamp_1 limit 1;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_2.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_2.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_2.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_2.q Tue Nov  5 07:01:32 2013
@@ -1,10 +1,12 @@
+set hive.fetch.task.conversion=more;
+
 drop table timestamp_2;
 
 create table timestamp_2 (t timestamp);
 alter table timestamp_2 set serde 'org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe';
 
-insert overwrite table timestamp_2 
-  select cast('2011-01-01 01:01:01' as timestamp) from src limit 1;
+insert overwrite table timestamp_2
+  select cast('2011-01-01 01:01:01' as timestamp) from src tablesample (1 rows);
 select cast(t as boolean) from timestamp_2 limit 1;
 select cast(t as tinyint) from timestamp_2 limit 1;
 select cast(t as smallint) from timestamp_2 limit 1;
@@ -15,7 +17,7 @@ select cast(t as double) from timestamp_
 select cast(t as string) from timestamp_2 limit 1;
 
 insert overwrite table timestamp_2
-  select '2011-01-01 01:01:01' from src limit 1;
+  select '2011-01-01 01:01:01' from src tablesample (1 rows);
 select cast(t as boolean) from timestamp_2 limit 1;
 select cast(t as tinyint) from timestamp_2 limit 1;
 select cast(t as smallint) from timestamp_2 limit 1;
@@ -26,7 +28,7 @@ select cast(t as double) from timestamp_
 select cast(t as string) from timestamp_2 limit 1;
 
 insert overwrite table timestamp_2
-  select '2011-01-01 01:01:01.1' from src limit 1;
+  select '2011-01-01 01:01:01.1' from src tablesample (1 rows);
 select cast(t as boolean) from timestamp_2 limit 1;
 select cast(t as tinyint) from timestamp_2 limit 1;
 select cast(t as smallint) from timestamp_2 limit 1;
@@ -37,7 +39,7 @@ select cast(t as double) from timestamp_
 select cast(t as string) from timestamp_2 limit 1;
 
 insert overwrite table timestamp_2
-  select '2011-01-01 01:01:01.0001' from src limit 1;
+  select '2011-01-01 01:01:01.0001' from src tablesample (1 rows);
 select cast(t as boolean) from timestamp_2 limit 1;
 select cast(t as tinyint) from timestamp_2 limit 1;
 select cast(t as smallint) from timestamp_2 limit 1;
@@ -48,7 +50,7 @@ select cast(t as double) from timestamp_
 select cast(t as string) from timestamp_2 limit 1;
 
 insert overwrite table timestamp_2
-  select '2011-01-01 01:01:01.000100000' from src limit 1;
+  select '2011-01-01 01:01:01.000100000' from src tablesample (1 rows);
 select cast(t as boolean) from timestamp_2 limit 1;
 select cast(t as tinyint) from timestamp_2 limit 1;
 select cast(t as smallint) from timestamp_2 limit 1;
@@ -59,7 +61,7 @@ select cast(t as double) from timestamp_
 select cast(t as string) from timestamp_2 limit 1;
 
 insert overwrite table timestamp_2
-  select '2011-01-01 01:01:01.001000011' from src limit 1;
+  select '2011-01-01 01:01:01.001000011' from src tablesample (1 rows);
 select cast(t as boolean) from timestamp_2 limit 1;
 select cast(t as tinyint) from timestamp_2 limit 1;
 select cast(t as smallint) from timestamp_2 limit 1;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_3.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_3.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_3.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_3.q Tue Nov  5 07:01:32 2013
@@ -1,10 +1,12 @@
+set hive.fetch.task.conversion=more;
+
 drop table timestamp_3;
 
 create table timestamp_3 (t timestamp);
 alter table timestamp_3 set serde 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe';
 
-insert overwrite table timestamp_3 
-  select cast(cast('1.3041352164485E9' as double) as timestamp) from src limit 1;
+insert overwrite table timestamp_3
+  select cast(cast('1.3041352164485E9' as double) as timestamp) from src tablesample (1 rows);
 select cast(t as boolean) from timestamp_3 limit 1;
 select cast(t as tinyint) from timestamp_3 limit 1;
 select cast(t as smallint) from timestamp_3 limit 1;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_comparison.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_comparison.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_comparison.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_comparison.q Tue Nov  5 07:01:32 2013
@@ -1,5 +1,6 @@
+set hive.fetch.task.conversion=more;
 
-select cast('2011-05-06 07:08:09' as timestamp) > 
+select cast('2011-05-06 07:08:09' as timestamp) >
   cast('2011-05-06 07:08:09' as timestamp) from src limit 1;
 
 select cast('2011-05-06 07:08:09' as timestamp) <

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_lazy.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_lazy.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_lazy.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_lazy.q Tue Nov  5 07:01:32 2013
@@ -1,6 +1,6 @@
 drop table timestamp_lazy;
 create table timestamp_lazy (t timestamp, key string, value string);
-insert overwrite table timestamp_lazy select cast('2011-01-01 01:01:01' as timestamp), key, value from src limit 5;
+insert overwrite table timestamp_lazy select cast('2011-01-01 01:01:01' as timestamp), key, value from src tablesample (5 rows);
 
 select t,key,value from timestamp_lazy ORDER BY key ASC, value ASC;
 select t,key,value from timestamp_lazy distribute by t sort by key ASC, value ASC;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_null.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_null.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_null.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_null.q Tue Nov  5 07:01:32 2013
@@ -1,6 +1,6 @@
 DROP TABLE IF EXISTS timestamp_null;
 CREATE TABLE timestamp_null (t1 TIMESTAMP);
-LOAD DATA LOCAL INPATH '../data/files/test.dat' OVERWRITE INTO TABLE timestamp_null;
+LOAD DATA LOCAL INPATH '../../data/files/test.dat' OVERWRITE INTO TABLE timestamp_null;
 
 SELECT * FROM timestamp_null LIMIT 1;
 

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_udf.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_udf.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_udf.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/timestamp_udf.q Tue Nov  5 07:01:32 2013
@@ -1,13 +1,15 @@
+set hive.fetch.task.conversion=more;
+
 drop table timestamp_udf;
 drop table timestamp_udf_string;
 
 create table timestamp_udf (t timestamp);
 create table timestamp_udf_string (t string);
-from src
+from (select * from src tablesample (1 rows)) s
   insert overwrite table timestamp_udf 
-    select '2011-05-06 07:08:09.1234567' limit 1
+    select '2011-05-06 07:08:09.1234567'
   insert overwrite table timestamp_udf_string
-    select '2011-05-06 07:08:09.1234567' limit 1;
+    select '2011-05-06 07:08:09.1234567';
 
 -- Test UDFs with Timestamp input
 select unix_timestamp(t), year(t), month(t), day(t), dayofmonth(t),

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/transform1.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/transform1.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/transform1.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/transform1.q Tue Nov  5 07:01:32 2013
@@ -12,7 +12,7 @@ SELECT transform(*) USING 'cat' AS (col 
 create table transform1_t2(col array<int>);
 
 insert overwrite table transform1_t2
-select array(1,2,3) from src limit 1;
+select array(1,2,3) from src tablesample (1 rows);
 
 EXPLAIN
 SELECT transform('0\0021\0022') USING 'cat' AS (col array<int>) FROM transform1_t2;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/truncate_column.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/truncate_column.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/truncate_column.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/truncate_column.q Tue Nov  5 07:01:32 2013
@@ -5,7 +5,7 @@ ROW FORMAT SERDE 'org.apache.hadoop.hive
 
 set hive.stats.autogather=true;
 
-INSERT OVERWRITE TABLE test_tab SELECT * FROM src LIMIT 10;
+INSERT OVERWRITE TABLE test_tab SELECT * FROM src tablesample (10 rows);
 
 DESC FORMATTED test_tab;
 
@@ -20,7 +20,7 @@ DESC FORMATTED test_tab;
 SELECT * FROM test_tab ORDER BY value;
 
 -- Truncate multiple columns
-INSERT OVERWRITE TABLE test_tab SELECT * FROM src LIMIT 10;
+INSERT OVERWRITE TABLE test_tab SELECT * FROM src tablesample (10 rows);
 
 TRUNCATE TABLE test_tab COLUMNS (key, value);
 
@@ -40,7 +40,7 @@ SELECT * FROM test_tab ORDER BY value;
 -- Test truncating with a binary serde
 ALTER TABLE test_tab SET SERDE 'org.apache.hadoop.hive.serde2.columnar.LazyBinaryColumnarSerDe';
 
-INSERT OVERWRITE TABLE test_tab SELECT * FROM src LIMIT 10;
+INSERT OVERWRITE TABLE test_tab SELECT * FROM src tablesample (10 rows);
 
 DESC FORMATTED test_tab;
 
@@ -65,7 +65,7 @@ SELECT * FROM test_tab ORDER BY value;
 -- Test truncating a partition
 CREATE TABLE test_tab_part (key STRING, value STRING) PARTITIONED BY (part STRING) STORED AS RCFILE;
 
-INSERT OVERWRITE TABLE test_tab_part PARTITION (part = '1') SELECT * FROM src LIMIT 10;
+INSERT OVERWRITE TABLE test_tab_part PARTITION (part = '1') SELECT * FROM src tablesample (10 rows);
 
 DESC FORMATTED test_tab_part PARTITION (part = '1');
 

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/truncate_column_merge.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/truncate_column_merge.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/truncate_column_merge.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/truncate_column_merge.q Tue Nov  5 07:01:32 2013
@@ -2,9 +2,9 @@
 
 CREATE TABLE test_tab (key STRING, value STRING) STORED AS RCFILE;
 
-INSERT OVERWRITE TABLE test_tab SELECT * FROM src LIMIT 5;
+INSERT OVERWRITE TABLE test_tab SELECT * FROM src tablesample (5 rows);
 
-INSERT INTO TABLE test_tab SELECT * FROM src LIMIT 5;
+INSERT INTO TABLE test_tab SELECT * FROM src tablesample (5 rows);
 
 -- The value should be 2 indicating the table has 2 files
 SELECT COUNT(DISTINCT INPUT__FILE__NAME) FROM test_tab;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/truncate_table.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/truncate_table.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/truncate_table.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/truncate_table.q Tue Nov  5 07:01:32 2013
@@ -1,5 +1,5 @@
 create table src_truncate (key string, value string);
-load data local inpath '../data/files/kv1.txt' into table src_truncate;;
+load data local inpath '../../data/files/kv1.txt' into table src_truncate;;
 
 create table srcpart_truncate (key string, value string) partitioned by (ds string, hr string);
 alter table srcpart_truncate add partition (ds='2008-04-08', hr='11');        
@@ -7,10 +7,10 @@ alter table srcpart_truncate add partiti
 alter table srcpart_truncate add partition (ds='2008-04-09', hr='11');
 alter table srcpart_truncate add partition (ds='2008-04-09', hr='12');
 
-load data local inpath '../data/files/kv1.txt' into table srcpart_truncate partition (ds='2008-04-08', hr='11');
-load data local inpath '../data/files/kv1.txt' into table srcpart_truncate partition (ds='2008-04-08', hr='12');
-load data local inpath '../data/files/kv1.txt' into table srcpart_truncate partition (ds='2008-04-09', hr='11');
-load data local inpath '../data/files/kv1.txt' into table srcpart_truncate partition (ds='2008-04-09', hr='12');
+load data local inpath '../../data/files/kv1.txt' into table srcpart_truncate partition (ds='2008-04-08', hr='11');
+load data local inpath '../../data/files/kv1.txt' into table srcpart_truncate partition (ds='2008-04-08', hr='12');
+load data local inpath '../../data/files/kv1.txt' into table srcpart_truncate partition (ds='2008-04-09', hr='11');
+load data local inpath '../../data/files/kv1.txt' into table srcpart_truncate partition (ds='2008-04-09', hr='12');
 
 set hive.fetch.task.convertion=more;
 

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/type_cast_1.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/type_cast_1.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/type_cast_1.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/type_cast_1.q Tue Nov  5 07:01:32 2013
@@ -1,3 +1,5 @@
+set hive.fetch.task.conversion=more;
+
 EXPLAIN
 SELECT IF(false, 1, cast(2 as smallint)) + 3 FROM src LIMIT 1;
 

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/type_conversions_1.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/type_conversions_1.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/type_conversions_1.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/type_conversions_1.q Tue Nov  5 07:01:32 2013
@@ -1,3 +1,4 @@
+set hive.fetch.task.conversion=more;
 
 -- casting from null should yield null
 select

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/type_widening.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/type_widening.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/type_widening.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/type_widening.q Tue Nov  5 07:01:32 2013
@@ -1,3 +1,4 @@
+set hive.fetch.task.conversion=more;
 -- Check for int, bigint automatic type widening conversions in UDFs, UNIONS
 EXPLAIN SELECT COALESCE(0, 9223372036854775807) FROM src LIMIT 1;
 SELECT COALESCE(0, 9223372036854775807) FROM src LIMIT 1;

Modified: hive/branches/tez/ql/src/test/queries/clientpositive/udaf_context_ngrams.q
URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/queries/clientpositive/udaf_context_ngrams.q?rev=1538880&r1=1538879&r2=1538880&view=diff
==============================================================================
--- hive/branches/tez/ql/src/test/queries/clientpositive/udaf_context_ngrams.q (original)
+++ hive/branches/tez/ql/src/test/queries/clientpositive/udaf_context_ngrams.q Tue Nov  5 07:01:32 2013
@@ -1,5 +1,5 @@
 CREATE TABLE kafka (contents STRING);
-LOAD DATA LOCAL INPATH '../data/files/text-en.txt' INTO TABLE kafka;
+LOAD DATA LOCAL INPATH '../../data/files/text-en.txt' INTO TABLE kafka;
 set mapred.reduce.tasks=1;
 set hive.exec.reducers.max=1;