You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by nz...@apache.org on 2010/07/23 01:57:42 UTC
svn commit: r966909 [3/27] - in /hadoop/hive/trunk: ./ contrib/
data/warehouse/src/ hbase-handler/ hwi/ jdbc/ odbc/ ql/
ql/src/test/org/apache/hadoop/hive/ql/ ql/src/test/queries/clientnegative/
ql/src/test/queries/clientpositive/ ql/src/test/results/c...
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullgroup5.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullgroup5.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullgroup5.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullgroup5.q Thu Jul 22 23:57:29 2010
@@ -1,8 +1,8 @@
-DROP TABLE tstparttbl;
+
CREATE TABLE tstparttbl(KEY STRING, VALUE STRING) PARTITIONED BY(ds string) STORED AS TEXTFILE;
LOAD DATA LOCAL INPATH '../data/files/kv1.txt' INTO TABLE tstparttbl PARTITION (ds='2009-04-09');
-DROP TABLE tstparttbl2;
+
CREATE TABLE tstparttbl2(KEY STRING, VALUE STRING) PARTITIONED BY(ds string) STORED AS TEXTFILE;
LOAD DATA LOCAL INPATH '../data/files/kv1.txt' INTO TABLE tstparttbl2 PARTITION (ds='2009-04-09');
@@ -22,5 +22,5 @@ select u.* from
)u;
-DROP TABLE tstparttbl;
-DROP TABLE tstparttbl2;
+
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullinput.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullinput.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullinput.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullinput.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
create table tstnullinut(a string, b string);
select x.* from tstnullinut x;
select x.a, count(1) from tstnullinut x group by x.a;
-drop table tstnullinut;
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullinput2.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullinput2.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullinput2.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullinput2.q Thu Jul 22 23:57:29 2010
@@ -1,8 +1,8 @@
-drop table nulltbl;
+
create table nulltbl(key int) partitioned by (ds string);
select key from nulltbl where ds='101';
select count(1) from nulltbl where ds='101';
-drop table nulltbl;
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullscript.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullscript.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullscript.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullscript.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-DROP TABLE nullscript;
+
CREATE TABLE nullscript(KEY STRING, VALUE STRING) STORED AS TEXTFILE;
LOAD DATA LOCAL INPATH '../data/files/kv1.txt' INTO TABLE nullscript;
LOAD DATA LOCAL INPATH '../data/files/nullfile.txt' INTO TABLE nullscript;
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_vs_table_metadata.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_vs_table_metadata.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_vs_table_metadata.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_vs_table_metadata.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-drop table partition_vs_table;
+
create table partition_vs_table(key string, value string) partitioned by (ds string);
@@ -11,4 +11,3 @@ insert overwrite table partition_vs_tabl
select key, value, newcol from partition_vs_table
order by key;
-drop table partition_vs_table;
\ No newline at end of file
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-drop table partition_test_partitioned;
+
create table partition_test_partitioned(key string, value string) partitioned by (dt string);
@@ -29,4 +29,4 @@ select key from partition_test_partition
select key from partition_test_partitioned;
select key from partition_test_partitioned where dt >=100 and dt <= 102;
-drop table partition_test_partitioned;
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat2.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat2.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat2.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat2.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-drop table partition_test_partitioned;
+
create table partition_test_partitioned(key string, value string) partitioned by (dt string);
@@ -9,4 +9,4 @@ alter table partition_test_partitioned s
insert overwrite table partition_test_partitioned partition(dt=102) select * from src1;
select * from partition_test_partitioned where dt >=100 and dt <= 102;
-drop table partition_test_partitioned;
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat3.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat3.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat3.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat3.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-drop table partition_test_partitioned;
+
create table partition_test_partitioned(key string, value string) partitioned by (dt string);
@@ -15,4 +15,4 @@ insert overwrite table partition_test_pa
show table extended like partition_test_partitioned partition(dt=101);
select key from partition_test_partitioned where dt=101;
-drop table partition_test_partitioned;
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/ppd_constant_expr.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/ppd_constant_expr.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/ppd_constant_expr.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/ppd_constant_expr.q Thu Jul 22 23:57:29 2010
@@ -10,4 +10,4 @@ INSERT OVERWRITE TABLE ppd_constant_expr
SELECT ppd_constant_expr.* FROM ppd_constant_expr;
-DROP TABLE ppd_constant_expr;
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/ppd_multi_insert.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/ppd_multi_insert.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/ppd_multi_insert.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/ppd_multi_insert.q Thu Jul 22 23:57:29 2010
@@ -1,7 +1,7 @@
set hive.optimize.ppd=true;
-DROP TABLE mi1;
-DROP TABLE mi2;
-DROP TABLE mi3;
+
+
+
CREATE TABLE mi1(key INT, value STRING) STORED AS TEXTFILE;
CREATE TABLE mi2(key INT, value STRING) STORED AS TEXTFILE;
CREATE TABLE mi3(key INT) PARTITIONED BY(ds STRING, hr STRING) STORED AS TEXTFILE;
@@ -24,6 +24,6 @@ SELECT mi2.* FROM mi2;
SELECT mi3.* FROM mi3;
dfs -cat ../build/ql/test/data/warehouse/mi4.out/*;
-DROP TABLE mi1;
-DROP TABLE mi2;
-DROP TABLE mi3;
+
+
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/progress_1.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/progress_1.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/progress_1.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/progress_1.q Thu Jul 22 23:57:29 2010
@@ -1,9 +1,9 @@
set hive.heartbeat.interval=5;
-DROP TABLE PROGRESS_1;
+
CREATE TABLE PROGRESS_1(key int, value string) STORED AS TEXTFILE;
LOAD DATA LOCAL INPATH '../data/files/kv6.txt' INTO TABLE PROGRESS_1;
select count(1) from PROGRESS_1 t1 join PROGRESS_1 t2 on t1.key=t2.key;
-DROP TABLE PROGRESS_1;
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/rand_partitionpruner2.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/rand_partitionpruner2.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/rand_partitionpruner2.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/rand_partitionpruner2.q Thu Jul 22 23:57:29 2010
@@ -1,5 +1,5 @@
-- scanning partitioned data
-drop table tmptable;
+
create table tmptable(key string, value string, hr string, ds string);
explain extended
@@ -11,4 +11,4 @@ insert overwrite table tmptable
select a.* from srcpart a where rand(1) < 0.1 and a.ds = '2008-04-08';
select * from tmptable x sort by x.key,x.value,x.ds,x.hr;
-drop table tmptable;
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_bigdata.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_bigdata.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_bigdata.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_bigdata.q Thu Jul 22 23:57:29 2010
@@ -1,7 +1,7 @@
set hive.map.aggr.hash.percentmemory = 0.3;
set hive.mapred.local.mem = 256;
-DROP TABLE columnTable_Bigdata;
+
CREATE table columnTable_Bigdata (key STRING, value STRING)
ROW FORMAT SERDE
'org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe'
@@ -15,4 +15,4 @@ INSERT OVERWRITE TABLE columnTable_Bigda
describe columnTable_Bigdata;
select count(columnTable_Bigdata.key) from columnTable_Bigdata;
-DROP TABLE columnTable_Bigdata;
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_columnar.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_columnar.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_columnar.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_columnar.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-DROP TABLE columnTable;
+
CREATE table columnTable (key STRING, value STRING)
ROW FORMAT SERDE
'org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe'
@@ -12,4 +12,4 @@ describe columnTable;
SELECT columnTable.* FROM columnTable;
-DROP TABLE columnTable;
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_default_format.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_default_format.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_default_format.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_default_format.q Thu Jul 22 23:57:29 2010
@@ -14,7 +14,6 @@ SET hive.default.fileformat = TextFile;
CREATE TABLE textfile_default_format_ctas AS SELECT key,value FROM rcfile_default_format_ctas;
DESCRIBE EXTENDED textfile_default_format_ctas;
-DROP TABLE rcfile_default_format;
-DROP TABLE rcfile_default_format_ctas;
-DROP TABLE rcfile_default_format_txtfile;
-DROP TABLE textfile_default_format_ctas;
\ No newline at end of file
+
+
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_lazydecompress.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_lazydecompress.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_lazydecompress.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_lazydecompress.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-DROP TABLE rcfileTableLazyDecompress;
+
CREATE table rcfileTableLazyDecompress (key STRING, value STRING) STORED AS RCFile;
FROM src
@@ -25,4 +25,3 @@ SELECT key, count(1) FROM rcfileTableLaz
set mapred.output.compress=false;
set hive.exec.compress.output=false;
-DROP TABLE rcfileTableLazyDecompress;
\ No newline at end of file
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_null_value.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_null_value.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_null_value.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_null_value.q Thu Jul 22 23:57:29 2010
@@ -1,7 +1,7 @@
CREATE TABLE src1_rc(key STRING, value STRING) STORED AS RCFILE;
INSERT OVERWRITE TABLE src1_rc SELECT * FROM src1;
SELECT * FROM src1_rc;
-DROP TABLE src1_rc;
+
CREATE TABLE dest1_rc(c1 INT, c2 STRING, c3 INT, c4 STRING) STORED AS RCFILE;
@@ -36,5 +36,5 @@ INSERT OVERWRITE TABLE dest1_rc SELECT c
SELECT dest1_rc.* FROM dest1_rc;
-DROP TABLE dest1_rc;
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_union.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_union.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_union.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_union.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-DROP TABLE rcfile_unionTable;
+
CREATE table rcfile_unionTable (b STRING, c STRING)
ROW FORMAT SERDE
'org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe'
@@ -12,4 +12,3 @@ SELECT b AS cola FROM rcfile_unionTable
UNION ALL
SELECT c AS cola FROM rcfile_unionTable) s;
-DROP TABLE rcfile_unionTable;
\ No newline at end of file
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/reduce_deduplicate.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/reduce_deduplicate.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/reduce_deduplicate.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/reduce_deduplicate.q Thu Jul 22 23:57:29 2010
@@ -2,7 +2,7 @@ set hive.enforce.bucketing = true;
set hive.exec.reducers.max = 1;
set hive.exec.script.trust = true;
-drop table bucket5_1;
+
CREATE TABLE bucket5_1(key string, value string) CLUSTERED BY (key) INTO 2 BUCKETS;
explain extended
@@ -15,10 +15,10 @@ select * from src cluster by key;
select sum(hash(key)),sum(hash(value)) from bucket5_1;
select sum(hash(key)),sum(hash(value)) from src;
-drop table complex_tbl_1;
+
create table complex_tbl_1(aid string, bid string, t int, ctime string, etime bigint, l string, et string) partitioned by (ds string);
-drop table complex_tbl_2;
+
create table complex_tbl_2(aet string, aes string) partitioned by (ds string);
explain extended
@@ -37,7 +37,7 @@ select s2.* from
)s
)s2;
-drop table complex_tbl_2;
-drop table complex_tbl_1;
-drop table bucket5_1;
+
+
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/rename_column.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/rename_column.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/rename_column.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/rename_column.q Thu Jul 22 23:57:29 2010
@@ -22,5 +22,5 @@ DESCRIBE kv_rename_test;
ALTER TABLE kv_rename_test CHANGE COLUMN a2 a INT AFTER b;
DESCRIBE kv_rename_test;
-DROP TABLE kv_rename_test;
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/repair.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/repair.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/repair.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/repair.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-DROP TABLE repairtable;
+
CREATE TABLE repairtable(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
@@ -13,4 +13,4 @@ MSCK REPAIR TABLE repairtable;
MSCK TABLE repairtable;
-DROP TABLE repairtable;
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/sample10.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/sample10.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/sample10.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/sample10.q Thu Jul 22 23:57:29 2010
@@ -21,4 +21,4 @@ select ds, count(1) from srcpartbucket t
select * from srcpartbucket where ds is not null;
-drop table srcpartbucket;
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/sample6.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/sample6.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/sample6.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/sample6.q Thu Jul 22 23:57:29 2010
@@ -46,5 +46,5 @@ ORDER BY key, value;
SELECT s.* FROM empty_bucket TABLESAMPLE (BUCKET 1 OUT OF 2 on key) s
ORDER BY key, value;
-drop table empty_bucket;
-drop table dest1;
+
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/semijoin.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/semijoin.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/semijoin.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/semijoin.q Thu Jul 22 23:57:29 2010
@@ -1,7 +1,7 @@
-drop table t1;
-drop table t2;
-drop table t3;
-drop table t4;
+
+
+
+
create table t1 as select cast(key as int) key, value from src where key <= 10;
@@ -77,7 +77,7 @@ select a.key from t3 a left semi join t1
explain select a.key from t3 a left semi join t2 b on a.key = b.key left outer join t1 c on a.value = c.value sort by a.key;
select a.key from t3 a left semi join t2 b on a.key = b.key left outer join t1 c on a.value = c.value sort by a.key;
-drop table t1;
-drop table t2;
-drop table t3;
-drop table t4;
+
+
+
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/show_tables.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/show_tables.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/show_tables.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/show_tables.q Thu Jul 22 23:57:29 2010
@@ -11,5 +11,5 @@ SHOW TABLES 'shtb_test1|shtb_test2';
SHOW TABLES 'shtb_test1|shtb_test2';
-DROP TABLE shtb_test1;
-DROP TABLE shtb_test2;
+
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/skewjoin.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/skewjoin.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/skewjoin.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/skewjoin.q Thu Jul 22 23:57:29 2010
@@ -1,11 +1,11 @@
set hive.optimize.skewjoin = true;
set hive.skewjoin.key = 2;
-DROP TABLE T1;
-DROP TABLE T2;
-DROP TABLE T3;
-DROP TABLE T4;
-DROP TABLE dest_j1;
+
+
+
+
+
CREATE TABLE T1(key STRING, val STRING) STORED AS TEXTFILE;
CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;
@@ -131,8 +131,7 @@ SELECT sum(hash(src1.key)), sum(hash(src
select /*+ mapjoin(v)*/ sum(hash(k.key)), sum(hash(v.val)) from T1 k left outer join T1 v on k.key+1=v.key;
-DROP TABLE dest_j1;
-DROP TABLE T1;
-DROP TABLE T2;
-DROP TABLE T3;
-DROP TABLE T4;
\ No newline at end of file
+
+
+
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_1.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_1.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_1.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_1.q Thu Jul 22 23:57:29 2010
@@ -1,6 +1,6 @@
-drop table smb_bucket_3;
-drop table smb_bucket_2;
-drop table smb_bucket_1;
+
+
+
create table smb_bucket_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
create table smb_bucket_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
@@ -47,6 +47,6 @@ select /*+mapjoin(b)*/ * from smb_bucket
select /*+mapjoin(b)*/ * from smb_bucket_1 a full outer join smb_bucket_2 b on a.key = b.key;
-drop table smb_bucket_3;
-drop table smb_bucket_2;
-drop table smb_bucket_1;
+
+
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_2.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_2.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_2.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_2.q Thu Jul 22 23:57:29 2010
@@ -1,6 +1,6 @@
-drop table smb_bucket_3;
-drop table smb_bucket_2;
-drop table smb_bucket_1;
+
+
+
create table smb_bucket_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
create table smb_bucket_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
@@ -47,6 +47,6 @@ select /*+mapjoin(b)*/ * from smb_bucket
select /*+mapjoin(b)*/ * from smb_bucket_1 a full outer join smb_bucket_3 b on a.key = b.key;
-drop table smb_bucket_3;
-drop table smb_bucket_2;
-drop table smb_bucket_1;
+
+
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_3.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_3.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_3.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_3.q Thu Jul 22 23:57:29 2010
@@ -1,6 +1,6 @@
-drop table smb_bucket_3;
-drop table smb_bucket_2;
-drop table smb_bucket_1;
+
+
+
create table smb_bucket_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
create table smb_bucket_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
@@ -47,6 +47,6 @@ select /*+mapjoin(b)*/ * from smb_bucket
select /*+mapjoin(b)*/ * from smb_bucket_2 a full outer join smb_bucket_3 b on a.key = b.key;
-drop table smb_bucket_3;
-drop table smb_bucket_2;
-drop table smb_bucket_1;
+
+
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_4.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_4.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_4.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_4.q Thu Jul 22 23:57:29 2010
@@ -1,6 +1,6 @@
-drop table smb_bucket_3;
-drop table smb_bucket_2;
-drop table smb_bucket_1;
+
+
+
create table smb_bucket_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
create table smb_bucket_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
@@ -66,6 +66,6 @@ select /*+mapjoin(a,b)*/ * from smb_buck
select /*+mapjoin(a,b)*/ * from smb_bucket_1 a full outer join smb_bucket_2 b on a.key = b.key full outer join smb_bucket_3 c on b.key=c.key;
-drop table smb_bucket_3;
-drop table smb_bucket_2;
-drop table smb_bucket_1;
+
+
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_5.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_5.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_5.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_5.q Thu Jul 22 23:57:29 2010
@@ -1,6 +1,6 @@
-drop table smb_bucket_3;
-drop table smb_bucket_2;
-drop table smb_bucket_1;
+
+
+
create table smb_bucket_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
create table smb_bucket_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE;
@@ -66,6 +66,6 @@ select /*+mapjoin(a,c)*/ * from smb_buck
select /*+mapjoin(a,c)*/ * from smb_bucket_1 a full outer join smb_bucket_2 b on a.key = b.key full outer join smb_bucket_3 c on b.key=c.key;
-drop table smb_bucket_3;
-drop table smb_bucket_2;
-drop table smb_bucket_1;
+
+
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_6.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_6.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_6.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_6.q Thu Jul 22 23:57:29 2010
@@ -2,10 +2,10 @@ set hive.enforce.bucketing = true;
set hive.enforce.sorting = true;
set hive.exec.reducers.max = 1;
-drop table smb_bucket4_1;
+
CREATE TABLE smb_bucket4_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS RCFILE;
-drop table smb_bucket4_2;
+
CREATE TABLE smb_bucket4_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS RCFILE;
create table smb_join_results(k1 int, v1 string, k2 int, v2 string);
@@ -71,7 +71,7 @@ explain
select /*+mapjoin(b,c)*/ * from smb_bucket4_1 a join smb_bucket4_2 b on a.key = b.key join smb_bucket4_2 c on b.key = c.key where a.key>1000;
select /*+mapjoin(b,c)*/ * from smb_bucket4_1 a join smb_bucket4_2 b on a.key = b.key join smb_bucket4_2 c on b.key = c.key where a.key>1000;
-drop table smb_join_results;
-drop table normal_join_results;
-drop table smb_bucket4_1;
-drop table smb_bucket4_2;
+
+
+
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_7.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_7.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_7.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_7.q Thu Jul 22 23:57:29 2010
@@ -2,15 +2,15 @@ set hive.enforce.bucketing = true;
set hive.enforce.sorting = true;
set hive.exec.reducers.max = 1;
-drop table smb_bucket4_1;
+
CREATE TABLE smb_bucket4_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS;
-drop table smb_bucket4_2;
+
CREATE TABLE smb_bucket4_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS;
-drop table smb_join_results;
-drop table smb_join_results_empty_bigtable;
-drop table normal_join_results;
+
+
+
create table smb_join_results(k1 int, v1 string, k2 int, v2 string);
create table smb_join_results_empty_bigtable(k1 int, v1 string, k2 int, v2 string);
create table normal_join_results(k1 int, v1 string, k2 int, v2 string);
@@ -47,8 +47,8 @@ select sum(hash(k1)) as k1, sum(hash(k2)
select sum(hash(k1)) as k1, sum(hash(k2)) as k2, sum(hash(v1)) as v1, sum(hash(v2)) as v2 from smb_join_results;
select sum(hash(k1)) as k1, sum(hash(k2)) as k2, sum(hash(v1)) as v1, sum(hash(v2)) as v2 from smb_join_results_empty_bigtable;
-drop table smb_join_results;
-drop table smb_join_results_empty_bigtable;
-drop table normal_join_results;
-drop table smb_bucket4_1;
-drop table smb_bucket4_2;
+
+
+
+
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_8.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_8.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_8.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_8.q Thu Jul 22 23:57:29 2010
@@ -1,7 +1,7 @@
set hive.enforce.bucketing = true;
set hive.exec.reducers.max = 1;
-drop table smb_bucket_input;
+
create table smb_bucket_input (key int, value string) stored as rcfile;
load data local inpath '../data/files/smb_bucket_input.rc' into table smb_bucket_input;
@@ -10,11 +10,11 @@ set hive.optimize.bucketmapjoin.sortedme
set hive.enforce.sorting = true;
set hive.exec.reducers.max = 1;
-drop table smb_bucket4_1;
+
CREATE TABLE smb_bucket4_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS;
-drop table smb_bucket4_2;
+
CREATE TABLE smb_bucket4_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS;
-drop table smb_bucket4_3;
+
CREATE TABLE smb_bucket4_3(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS;
insert overwrite table smb_bucket4_1 select * from smb_bucket_input where key=4 or key=2000 or key=4000;
@@ -84,7 +84,6 @@ insert overwrite table smb_bucket4_3 sel
select /*+mapjoin(b,c)*/ * from smb_bucket4_1 a full outer join smb_bucket4_2 b on a.key = b.key
full outer join smb_bucket4_3 c on a.key=c.key;
-drop table smb_bucket4_1;
-drop table smb_bucket4_2;
-drop table smb_bucket4_3;
-drop table smb_bucket_input;
\ No newline at end of file
+
+
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/symlink_text_input_format.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/symlink_text_input_format.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/symlink_text_input_format.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/symlink_text_input_format.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-DROP TABLE symlink_text_input_format;
+
EXPLAIN
CREATE TABLE symlink_text_input_format (key STRING, value STRING) STORED AS INPUTFORMAT 'org.apache.hadoop.hive.ql.io.SymlinkTextInputFormat' OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat';
@@ -20,5 +20,5 @@ EXPLAIN SELECT count(1) FROM symlink_tex
SELECT count(1) FROM symlink_text_input_format;
-DROP TABLE symlink_text_input_format;
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/tablename_with_select.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/tablename_with_select.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/tablename_with_select.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/tablename_with_select.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-DROP TABLE tmp_select;
+
CREATE TABLE tmp_select(a INT, b STRING);
DESCRIBE tmp_select;
@@ -6,4 +6,4 @@ INSERT OVERWRITE TABLE tmp_select SELECT
SELECT a, b FROM tmp_select ORDER BY a;
-DROP TABLE tmp_select;
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/transform1.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/transform1.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/transform1.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/transform1.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-drop table transform1_t1;
+
create table transform1_t1(a string, b string);
EXPLAIN
@@ -6,9 +6,9 @@ SELECT transform(*) USING 'cat' AS (col
SELECT transform(*) USING 'cat' AS (col array<bigint>) FROM transform1_t1;
-drop table transform1_t1;
-drop table transform1_t2;
+
+
create table transform1_t2(col array<int>);
insert overwrite table transform1_t2
@@ -19,5 +19,5 @@ SELECT transform('0\0021\0022') USING 'c
SELECT transform('0\0021\0022') USING 'cat' AS (col array<int>) FROM transform1_t2;
-drop table transform1_t2;
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_concat_insert1.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_concat_insert1.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_concat_insert1.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_concat_insert1.q Thu Jul 22 23:57:29 2010
@@ -5,4 +5,3 @@ INSERT OVERWRITE TABLE dest1 SELECT '123
SELECT dest1.* FROM dest1;
-DROP TABLE dest1;
\ No newline at end of file
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_concat_insert2.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_concat_insert2.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_concat_insert2.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_concat_insert2.q Thu Jul 22 23:57:29 2010
@@ -5,4 +5,4 @@ INSERT OVERWRITE TABLE dest1 SELECT conc
SELECT dest1.* FROM dest1;
-DROP TABLE dest1;
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_field.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_field.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_field.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_field.q Thu Jul 22 23:57:29 2010
@@ -23,7 +23,7 @@ SELECT
field(4, 1, 2, NULL, 4)
FROM src LIMIT 1;
-DROP TABLE test_table;
+
CREATE TABLE test_table(col1 STRING, col2 STRING) STORED AS TEXTFILE;
LOAD DATA LOCAL INPATH '../data/files/kv1.txt' INTO TABLE test_table;
@@ -40,7 +40,7 @@ select col1,col2,
field(col1, col2, "66")
from test_table where col1="86" or col1="66";
-DROP TABLE test_table1;
+
CREATE TABLE test_table1(col1 int, col2 string) STORED AS TEXTFILE;
LOAD DATA LOCAL INPATH '../data/files/kv1.txt' INTO TABLE test_table1;
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_length.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_length.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_length.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_length.q Thu Jul 22 23:57:29 2010
@@ -12,4 +12,3 @@ CREATE TABLE dest1(name STRING) STORED A
LOAD DATA LOCAL INPATH '../data/files/kv4.txt' INTO TABLE dest1;
EXPLAIN SELECT length(dest1.name) FROM dest1;
SELECT length(dest1.name) FROM dest1;
-DROP TABLE dest1;
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_reverse.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_reverse.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_reverse.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_reverse.q Thu Jul 22 23:57:29 2010
@@ -13,4 +13,3 @@ DROP TABLE dest1;
CREATE TABLE dest1(name STRING) STORED AS TEXTFILE;
LOAD DATA LOCAL INPATH '../data/files/kv4.txt' INTO TABLE dest1;
SELECT count(1) FROM dest1 WHERE reverse(dest1.name) = _UTF-8 0xE993AEE982B5;
-DROP TABLE dest1;
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_sentences.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_sentences.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_sentences.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_sentences.q Thu Jul 22 23:57:29 2010
@@ -5,10 +5,10 @@ SELECT explode(sentences(unhex("48697665
INSERT OVERWRITE TABLE sent_tmp2
SELECT explode(val) AS val FROM sent_tmp;
SELECT hex(val) FROM sent_tmp2;
+
DROP TABLE sent_tmp;
DROP TABLE sent_tmp2;
-
CREATE TABLE sent_tmp (val array<string>);
CREATE TABLE sent_tmp2 (val string);
INSERT OVERWRITE TABLE sent_tmp
@@ -16,7 +16,7 @@ SELECT explode(sentences(unhex("48697665
INSERT OVERWRITE TABLE sent_tmp2
SELECT explode(val) AS val FROM sent_tmp;
SELECT hex(val) FROM sent_tmp2;
-DROP TABLE sent_tmp;
-DROP TABLE sent_tmp2;
+
+
SELECT sentences("Hive is an excellent tool for data querying\; and perhaps more versatile than machine translation!! Multiple, ill-formed sentences...confounding punctuation--and yet this UDF still works!!!!") FROM src LIMIT 1;
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/union10.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/union10.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/union10.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/union10.q Thu Jul 22 23:57:29 2010
@@ -2,7 +2,6 @@ set hive.map.aggr = true;
-- union case: all subqueries are a map-reduce jobs, 3 way union, same input for all sub-queries, followed by filesink
-drop table tmptable;
create table tmptable(key string, value int);
explain
@@ -24,5 +23,3 @@ insert overwrite table tmptable
select * from tmptable x sort by x.key;
-drop table tmptable;
-
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/union12.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/union12.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/union12.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/union12.q Thu Jul 22 23:57:29 2010
@@ -2,7 +2,6 @@ set hive.map.aggr = true;
-- union case: all subqueries are a map-reduce jobs, 3 way union, different inputs for all sub-queries, followed by filesink
-drop table tmptable;
create table tmptable(key string, value int);
explain
@@ -22,5 +21,3 @@ insert overwrite table tmptable
select 'tst3' as key, count(1) as value from srcbucket s3) unionsrc;
select * from tmptable x sort by x.key;
-
-drop table tmptable;
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/union17.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/union17.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/union17.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/union17.q Thu Jul 22 23:57:29 2010
@@ -1,6 +1,3 @@
-drop table DEST1;
-drop table DEST2;
-
CREATE TABLE DEST1(key STRING, value STRING) STORED AS TEXTFILE;
CREATE TABLE DEST2(key STRING, val1 STRING, val2 STRING) STORED AS TEXTFILE;
@@ -21,6 +18,3 @@ INSERT OVERWRITE TABLE DEST2 SELECT unio
SELECT DEST1.* FROM DEST1;
SELECT DEST2.* FROM DEST2;
-
-drop table DEST1;
-drop table DEST2;
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/union18.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/union18.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/union18.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/union18.q Thu Jul 22 23:57:29 2010
@@ -1,6 +1,3 @@
-drop table DEST1;
-drop table DEST2;
-
CREATE TABLE DEST1(key STRING, value STRING) STORED AS TEXTFILE;
CREATE TABLE DEST2(key STRING, val1 STRING, val2 STRING) STORED AS TEXTFILE;
@@ -21,6 +18,3 @@ INSERT OVERWRITE TABLE DEST2 SELECT unio
SELECT DEST1.* FROM DEST1 SORT BY DEST1.key, DEST1.value;
SELECT DEST2.* FROM DEST2 SORT BY DEST2.key, DEST2.val1, DEST2.val2;
-
-drop table DEST1;
-drop table DEST2;
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/union19.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/union19.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/union19.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/union19.q Thu Jul 22 23:57:29 2010
@@ -1,5 +1,5 @@
-drop table DEST1;
-drop table DEST2;
+
+
CREATE TABLE DEST1(key STRING, value STRING) STORED AS TEXTFILE;
CREATE TABLE DEST2(key STRING, val1 STRING, val2 STRING) STORED AS TEXTFILE;
@@ -22,5 +22,5 @@ INSERT OVERWRITE TABLE DEST2 SELECT unio
SELECT DEST1.* FROM DEST1 SORT BY DEST1.key, DEST1.value;
SELECT DEST2.* FROM DEST2 SORT BY DEST2.key, DEST2.val1, DEST2.val2;
-drop table DEST1;
-drop table DEST2;
+
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/union22.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/union22.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/union22.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/union22.q Thu Jul 22 23:57:29 2010
@@ -1,7 +1,7 @@
-drop table dst_union22;
+
create table dst_union22(k1 string, k2 string, k3 string, k4 string) partitioned by (ds string);
-drop table dst_union22_delta;
+
create table dst_union22_delta(k0 string, k1 string, k2 string, k3 string, k4 string, k5 string) partitioned by (ds string);
insert overwrite table dst_union22 partition (ds='1')
@@ -40,5 +40,5 @@ subq;
select * from dst_union22 where ds = '2' order by k1;
-drop table dst_union22;
-drop table dst_union22_delta;
+
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/union3.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/union3.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/union3.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/union3.q Thu Jul 22 23:57:29 2010
@@ -17,7 +17,7 @@ FROM (
FROM (SELECT * FROM src LIMIT 1) s2
) a;
-DROP TABLE union_out;
+
CREATE TABLE union_out (id int);
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/union4.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/union4.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/union4.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/union4.q Thu Jul 22 23:57:29 2010
@@ -2,7 +2,7 @@ set hive.map.aggr = true;
-- union case: both subqueries are map-reduce jobs on same input, followed by filesink
-drop table tmptable;
+
create table tmptable(key string, value int);
explain
@@ -18,4 +18,4 @@ select unionsrc.key, unionsrc.value FROM
select * from tmptable x sort by x.key;
-drop table tmptable;
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/union6.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/union6.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/union6.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/union6.q Thu Jul 22 23:57:29 2010
@@ -2,7 +2,7 @@ set hive.map.aggr = true;
-- union case: 1 subquery is a map-reduce job, different inputs for sub-queries, followed by filesink
-drop table tmptable;
+
create table tmptable(key string, value string);
explain
@@ -18,4 +18,4 @@ select unionsrc.key, unionsrc.value FROM
select * from tmptable x sort by x.key, x.value;
-drop table tmptable;
+
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/uniquejoin.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/uniquejoin.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/uniquejoin.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/uniquejoin.q Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-DROP TABLE T1;
-DROP TABLE T2;
-DROP TABLE T3;
-
CREATE TABLE T1(key STRING, val STRING) STORED AS TEXTFILE;
CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;
CREATE TABLE T3(key STRING, val STRING) STORED AS TEXTFILE;
@@ -27,7 +23,3 @@ SELECT a.key, b.key, c.key;
FROM UNIQUEJOIN PRESERVE T1 a (a.key), T2 b(b.key)
SELECT a.key, b.key;
-
-DROP TABLE T1;
-DROP TABLE T2;
-DROP TABLE T3;
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/addpart1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/addpart1.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/addpart1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/addpart1.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table addpart1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table addpart1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table addpart1 (a int) partitioned by (b string, c string)
PREHOOK: type: CREATETABLE
POSTHOOK: query: create table addpart1 (a int) partitioned by (b string, c string)
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/alter_non_native.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/alter_non_native.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/alter_non_native.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/alter_non_native.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE non_native1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE non_native1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE non_native1(key int, value string)
STORED BY 'org.apache.hadoop.hive.ql.metadata.DefaultStorageHandler'
PREHOOK: type: CREATETABLE
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/altern1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/altern1.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/altern1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/altern1.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table altern1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table altern1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table altern1(a int, b int) partitioned by (ds string)
PREHOOK: type: CREATETABLE
POSTHOOK: query: create table altern1(a int, b int) partitioned by (ds string)
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/create_insert_outputformat.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/create_insert_outputformat.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/create_insert_outputformat.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/create_insert_outputformat.q.out Thu Jul 22 23:57:29 2010
@@ -1,5 +1 @@
-PREHOOK: query: DROP TABLE table_test_output_fomat
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE table_test_output_fomat
-POSTHOOK: type: DROPTABLE
FAILED: Error in semantic analysis: Output Format must implement HiveOutputFormat, otherwise it should be either IgnoreKeyTextOutputFormat or SequenceFileOutputFormat
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/create_view_failure1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/create_view_failure1.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/create_view_failure1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/create_view_failure1.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE xxx12
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE xxx12
-POSTHOOK: type: DROPTABLE
PREHOOK: query: DROP VIEW xxx12
PREHOOK: type: DROPVIEW
POSTHOOK: query: DROP VIEW xxx12
@@ -15,6 +11,6 @@ POSTHOOK: type: CREATETABLE
POSTHOOK: Output: default@xxx12
PREHOOK: query: CREATE VIEW xxx12 AS SELECT key FROM src
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/Users/jsichi/open/hive-trunk/build/ql/tmp/194996627/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-43-17_176_344624644454180304/10000
FAILED: Error in metadata: AlreadyExistsException(message:Table xxx12 already exists)
FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/create_view_failure2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/create_view_failure2.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/create_view_failure2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/create_view_failure2.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE xxx4
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE xxx4
-POSTHOOK: type: DROPTABLE
PREHOOK: query: DROP VIEW xxx4
PREHOOK: type: DROPVIEW
POSTHOOK: query: DROP VIEW xxx4
@@ -9,11 +5,11 @@ POSTHOOK: type: DROPVIEW
PREHOOK: query: -- views and tables share the same namespace
CREATE VIEW xxx4 AS SELECT key FROM src
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/Users/jsichi/open/hive-trunk/build/ql/tmp/1327466632/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-43-17_460_630133670553073559/10000
POSTHOOK: query: -- views and tables share the same namespace
CREATE VIEW xxx4 AS SELECT key FROM src
POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/Users/jsichi/open/hive-trunk/build/ql/tmp/1327466632/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-43-17_460_630133670553073559/10000
POSTHOOK: Output: default@xxx4
PREHOOK: query: CREATE TABLE xxx4(key int)
PREHOOK: type: CREATETABLE
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/ctas.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/ctas.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/ctas.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/ctas.q.out Thu Jul 22 23:57:29 2010
@@ -1,5 +1 @@
-PREHOOK: query: drop table nzhang_ctas4
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table nzhang_ctas4
-POSTHOOK: type: DROPTABLE
FAILED: Error in semantic analysis: CREATE-TABLE-AS-SELECT cannot create external table.
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/ddltime.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/ddltime.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/ddltime.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/ddltime.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table T2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table T2
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table T2 like srcpart
PREHOOK: type: CREATETABLE
POSTHOOK: query: create table T2 like srcpart
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/deletejar.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/deletejar.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/deletejar.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/deletejar.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE DELETEJAR
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE DELETEJAR
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE DELETEJAR(KEY STRING, VALUE STRING) ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.TestSerDe' STORED AS TEXTFILE
PREHOOK: type: CREATETABLE
FAILED: Error in metadata: Cannot validate serde: org.apache.hadoop.hive.serde2.TestSerDe
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/drop_view_failure1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/drop_view_failure1.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/drop_view_failure1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/drop_view_failure1.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE xxx1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE xxx1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE xxx1(key int)
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE xxx1(key int)
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/dyn_part1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/dyn_part1.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/dyn_part1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/dyn_part1.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table dynamic_partition
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table dynamic_partition
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table dynamic_partition (key string) partitioned by (value string)
PREHOOK: type: CREATETABLE
POSTHOOK: query: create table dynamic_partition (key string) partitioned by (value string)
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/dyn_part2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/dyn_part2.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/dyn_part2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/dyn_part2.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table nzhang_part1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table nzhang_part1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table nzhang_part1 (key string, value string) partitioned by (ds string, hr string)
PREHOOK: type: CREATETABLE
POSTHOOK: query: create table nzhang_part1 (key string, value string) partitioned by (ds string, hr string)
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/external1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/external1.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/external1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/external1.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table external1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table external1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create external table external1(a int, b int) location 'invalidscheme://data.s3ndemo.hive/kv'
PREHOOK: type: CREATETABLE
FAILED: Error in metadata: MetaException(message:Got exception: java.io.IOException No FileSystem for scheme: invalidscheme)
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/external2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/external2.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/external2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/external2.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table external2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table external2
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create external table external2(a int, b int) partitioned by (ds string)
PREHOOK: type: CREATETABLE
POSTHOOK: query: create external table external2(a int, b int) partitioned by (ds string)
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_create_tbl1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_create_tbl1.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_create_tbl1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_create_tbl1.q.out Thu Jul 22 23:57:29 2010
@@ -1,5 +1 @@
-PREHOOK: query: DROP TABLE inv_valid_tbl1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE inv_valid_tbl1
-POSTHOOK: type: DROPTABLE
FAILED: Error in semantic analysis: DATE, DATETIME, and TIMESTAMP types aren't supported yet. Please use STRING instead.
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/invalidate_view1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/invalidate_view1.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/invalidate_view1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/invalidate_view1.q.out Thu Jul 22 23:57:29 2010
@@ -6,10 +6,6 @@ PREHOOK: query: DROP VIEW xxx9
PREHOOK: type: DROPVIEW
POSTHOOK: query: DROP VIEW xxx9
POSTHOOK: type: DROPVIEW
-PREHOOK: query: DROP TABLE xxx10
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE xxx10
-POSTHOOK: type: DROPTABLE
PREHOOK: query: -- create two levels of view reference, then invalidate intermediate view
-- by dropping a column from underlying table, and verify that
-- querying outermost view results in full error context
@@ -23,17 +19,17 @@ POSTHOOK: type: CREATETABLE
POSTHOOK: Output: default@xxx10
PREHOOK: query: CREATE VIEW xxx9 AS SELECT * FROM xxx10
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/Users/jsichi/open/hive-trunk/build/ql/tmp/1720348561/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-43-31_670_5934534815948379913/10000
POSTHOOK: query: CREATE VIEW xxx9 AS SELECT * FROM xxx10
POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/Users/jsichi/open/hive-trunk/build/ql/tmp/1720348561/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-43-31_670_5934534815948379913/10000
POSTHOOK: Output: default@xxx9
PREHOOK: query: CREATE VIEW xxx8 AS SELECT * FROM xxx9 xxx
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/Users/jsichi/open/hive-trunk/build/ql/tmp/1891844844/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-43-31_706_855643045189903710/10000
POSTHOOK: query: CREATE VIEW xxx8 AS SELECT * FROM xxx9 xxx
POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/Users/jsichi/open/hive-trunk/build/ql/tmp/1891844844/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-43-31_706_855643045189903710/10000
POSTHOOK: Output: default@xxx8
PREHOOK: query: ALTER TABLE xxx10 REPLACE COLUMNS (key int)
PREHOOK: type: ALTERTABLE_REPLACECOLS
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/load_non_native.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/load_non_native.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/load_non_native.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/load_non_native.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE non_native2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE non_native2
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE non_native2(key int, value string)
STORED BY 'org.apache.hadoop.hive.ql.metadata.DefaultStorageHandler'
PREHOOK: type: CREATETABLE
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat.q.out Thu Jul 22 23:57:29 2010
@@ -1,16 +1,14 @@
PREHOOK: query: -- test for loading into tables with the correct file format
-- test for loading into partitions with the correct file format
-DROP TABLE load_wrong_fileformat_T1
-PREHOOK: type: DROPTABLE
+
+CREATE TABLE load_wrong_fileformat_T1(name STRING) STORED AS SEQUENCEFILE
+PREHOOK: type: CREATETABLE
POSTHOOK: query: -- test for loading into tables with the correct file format
-- test for loading into partitions with the correct file format
-DROP TABLE load_wrong_fileformat_T1
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE load_wrong_fileformat_T1(name STRING) STORED AS SEQUENCEFILE
-PREHOOK: type: CREATETABLE
-POSTHOOK: query: CREATE TABLE load_wrong_fileformat_T1(name STRING) STORED AS SEQUENCEFILE
+
+CREATE TABLE load_wrong_fileformat_T1(name STRING) STORED AS SEQUENCEFILE
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: default@load_wrong_fileformat_T1
PREHOOK: query: LOAD DATA LOCAL INPATH '../data/files/kv1.txt' INTO TABLE load_wrong_fileformat_T1
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat_rc_seq.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat_rc_seq.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat_rc_seq.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat_rc_seq.q.out Thu Jul 22 23:57:29 2010
@@ -1,16 +1,14 @@
PREHOOK: query: -- test for loading into tables with the correct file format
-- test for loading into partitions with the correct file format
-DROP TABLE T1
-PREHOOK: type: DROPTABLE
+
+CREATE TABLE T1(name STRING) STORED AS RCFILE
+PREHOOK: type: CREATETABLE
POSTHOOK: query: -- test for loading into tables with the correct file format
-- test for loading into partitions with the correct file format
-DROP TABLE T1
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE T1(name STRING) STORED AS RCFILE
-PREHOOK: type: CREATETABLE
-POSTHOOK: query: CREATE TABLE T1(name STRING) STORED AS RCFILE
+
+CREATE TABLE T1(name STRING) STORED AS RCFILE
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: default@T1
PREHOOK: query: LOAD DATA LOCAL INPATH '../data/files/kv1.seq' INTO TABLE T1
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat_txt_seq.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat_txt_seq.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat_txt_seq.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat_txt_seq.q.out Thu Jul 22 23:57:29 2010
@@ -1,16 +1,14 @@
PREHOOK: query: -- test for loading into tables with the correct file format
-- test for loading into partitions with the correct file format
-DROP TABLE T1
-PREHOOK: type: DROPTABLE
+
+CREATE TABLE T1(name STRING) STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
POSTHOOK: query: -- test for loading into tables with the correct file format
-- test for loading into partitions with the correct file format
-DROP TABLE T1
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE T1(name STRING) STORED AS TEXTFILE
-PREHOOK: type: CREATETABLE
-POSTHOOK: query: CREATE TABLE T1(name STRING) STORED AS TEXTFILE
+
+CREATE TABLE T1(name STRING) STORED AS TEXTFILE
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: default@T1
PREHOOK: query: LOAD DATA LOCAL INPATH '../data/files/kv1.seq' INTO TABLE T1
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/nopart_insert.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/nopart_insert.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/nopart_insert.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/nopart_insert.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE nopart_insert
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE nopart_insert
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE nopart_insert(a STRING, b STRING) PARTITIONED BY (ds STRING)
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE nopart_insert(a STRING, b STRING) PARTITIONED BY (ds STRING)
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/nopart_load.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/nopart_load.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/nopart_load.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/nopart_load.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE nopart_load
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE nopart_load
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE nopart_load(a STRING, b STRING) PARTITIONED BY (ds STRING)
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE nopart_load(a STRING, b STRING) PARTITIONED BY (ds STRING)
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/smb_bucketmapjoin.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/smb_bucketmapjoin.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/smb_bucketmapjoin.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/smb_bucketmapjoin.q.out Thu Jul 22 23:57:29 2010
@@ -1,16 +1,8 @@
-PREHOOK: query: drop table smb_bucket4_1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table smb_bucket4_1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE smb_bucket4_1(key int, value string) CLUSTERED BY (key) INTO 2 BUCKETS
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE smb_bucket4_1(key int, value string) CLUSTERED BY (key) INTO 2 BUCKETS
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: default@smb_bucket4_1
-PREHOOK: query: drop table smb_bucket4_2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table smb_bucket4_2
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE smb_bucket4_2(key int, value string) CLUSTERED BY (key) INTO 2 BUCKETS
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE smb_bucket4_2(key int, value string) CLUSTERED BY (key) INTO 2 BUCKETS
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/union2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/union2.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/union2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/union2.q.out Thu Jul 22 23:57:29 2010
@@ -1,11 +1,3 @@
-PREHOOK: query: drop table union2_t1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table union2_t1
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: drop table union2_t2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table union2_t2
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table if not exists union2_t1(r string, c string, v string)
PREHOOK: type: CREATETABLE
POSTHOOK: query: create table if not exists union2_t1(r string, c string, v string)
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/add_part_exist.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/add_part_exist.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/add_part_exist.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/add_part_exist.q.out Thu Jul 22 23:57:29 2010
@@ -49,8 +49,3 @@ POSTHOOK: type: SHOWPARTITIONS
ds=2010-01-01
ds=2010-01-02
ds=2010-01-03
-PREHOOK: query: DROP TABLE add_part_test
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE add_part_test
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@add_part_test