You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by nz...@apache.org on 2010/07/23 01:57:42 UTC

svn commit: r966909 [3/27] - in /hadoop/hive/trunk: ./ contrib/ data/warehouse/src/ hbase-handler/ hwi/ jdbc/ odbc/ ql/ ql/src/test/org/apache/hadoop/hive/ql/ ql/src/test/queries/clientnegative/ ql/src/test/queries/clientpositive/ ql/src/test/results/c...

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullgroup5.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullgroup5.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullgroup5.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullgroup5.q Thu Jul 22 23:57:29 2010
@@ -1,8 +1,8 @@
-DROP TABLE tstparttbl;
+
 CREATE TABLE tstparttbl(KEY STRING, VALUE STRING) PARTITIONED BY(ds string) STORED AS TEXTFILE;
 LOAD DATA LOCAL INPATH '../data/files/kv1.txt' INTO TABLE tstparttbl PARTITION (ds='2009-04-09');
 
-DROP TABLE tstparttbl2;
+
 CREATE TABLE tstparttbl2(KEY STRING, VALUE STRING) PARTITIONED BY(ds string) STORED AS TEXTFILE;
 LOAD DATA LOCAL INPATH '../data/files/kv1.txt' INTO TABLE tstparttbl2 PARTITION (ds='2009-04-09');
 
@@ -22,5 +22,5 @@ select u.* from
 )u;
 
 
-DROP TABLE tstparttbl;
-DROP TABLE tstparttbl2;
+
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullinput.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullinput.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullinput.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullinput.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
 create table tstnullinut(a string, b string);
 select x.* from tstnullinut x;
 select x.a, count(1) from tstnullinut x group by x.a;
-drop table tstnullinut;
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullinput2.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullinput2.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullinput2.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullinput2.q Thu Jul 22 23:57:29 2010
@@ -1,8 +1,8 @@
-drop table nulltbl;
+
 
 create table nulltbl(key int) partitioned by (ds string);
 select key from nulltbl where ds='101';
 
 select count(1) from nulltbl where ds='101';
 
-drop table nulltbl;
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullscript.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullscript.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullscript.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/nullscript.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-DROP TABLE nullscript;
+
 CREATE TABLE nullscript(KEY STRING, VALUE STRING) STORED AS TEXTFILE;
 LOAD DATA LOCAL INPATH '../data/files/kv1.txt' INTO TABLE nullscript;
 LOAD DATA LOCAL INPATH '../data/files/nullfile.txt' INTO TABLE nullscript;

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_vs_table_metadata.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_vs_table_metadata.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_vs_table_metadata.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_vs_table_metadata.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-drop table partition_vs_table;
+
 
 create table partition_vs_table(key string, value string) partitioned by (ds string);
 
@@ -11,4 +11,3 @@ insert overwrite table partition_vs_tabl
 select key, value, newcol from partition_vs_table
 order by key;
 
-drop table partition_vs_table;
\ No newline at end of file

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-drop table partition_test_partitioned;
+
 
 create table partition_test_partitioned(key string, value string) partitioned by (dt string);
 
@@ -29,4 +29,4 @@ select key from partition_test_partition
 select key from partition_test_partitioned;
 
 select key from partition_test_partitioned where dt >=100 and dt <= 102;
-drop table partition_test_partitioned;
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat2.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat2.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat2.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat2.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-drop table partition_test_partitioned;
+
 
 create table partition_test_partitioned(key string, value string) partitioned by (dt string);
 
@@ -9,4 +9,4 @@ alter table partition_test_partitioned s
 insert overwrite table partition_test_partitioned partition(dt=102) select * from src1;
 
 select * from partition_test_partitioned where dt >=100 and dt <= 102;
-drop table partition_test_partitioned;
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat3.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat3.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat3.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/partition_wise_fileformat3.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-drop table partition_test_partitioned;
+
 
 create table partition_test_partitioned(key string, value string) partitioned by (dt string);
 
@@ -15,4 +15,4 @@ insert overwrite table partition_test_pa
 show table extended like partition_test_partitioned partition(dt=101);
 select key from partition_test_partitioned where dt=101;
 
-drop table partition_test_partitioned;
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/ppd_constant_expr.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/ppd_constant_expr.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/ppd_constant_expr.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/ppd_constant_expr.q Thu Jul 22 23:57:29 2010
@@ -10,4 +10,4 @@ INSERT OVERWRITE TABLE ppd_constant_expr
 
 SELECT ppd_constant_expr.* FROM ppd_constant_expr;
 
-DROP TABLE ppd_constant_expr;
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/ppd_multi_insert.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/ppd_multi_insert.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/ppd_multi_insert.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/ppd_multi_insert.q Thu Jul 22 23:57:29 2010
@@ -1,7 +1,7 @@
 set hive.optimize.ppd=true;
-DROP TABLE mi1;
-DROP TABLE mi2;
-DROP TABLE mi3;
+
+
+
 CREATE TABLE mi1(key INT, value STRING) STORED AS TEXTFILE;
 CREATE TABLE mi2(key INT, value STRING) STORED AS TEXTFILE;
 CREATE TABLE mi3(key INT) PARTITIONED BY(ds STRING, hr STRING) STORED AS TEXTFILE;
@@ -24,6 +24,6 @@ SELECT mi2.* FROM mi2;
 SELECT mi3.* FROM mi3;
 dfs -cat ../build/ql/test/data/warehouse/mi4.out/*;
 
-DROP TABLE mi1;
-DROP TABLE mi2;
-DROP TABLE mi3;
+
+
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/progress_1.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/progress_1.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/progress_1.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/progress_1.q Thu Jul 22 23:57:29 2010
@@ -1,9 +1,9 @@
 set hive.heartbeat.interval=5; 
 
-DROP TABLE PROGRESS_1;
+
 CREATE TABLE PROGRESS_1(key int, value string) STORED AS TEXTFILE;
 LOAD DATA LOCAL INPATH '../data/files/kv6.txt' INTO TABLE PROGRESS_1;
 
 select count(1) from PROGRESS_1 t1 join PROGRESS_1 t2 on t1.key=t2.key;
 
-DROP TABLE PROGRESS_1;
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/rand_partitionpruner2.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/rand_partitionpruner2.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/rand_partitionpruner2.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/rand_partitionpruner2.q Thu Jul 22 23:57:29 2010
@@ -1,5 +1,5 @@
 -- scanning partitioned data
-drop table tmptable;
+
 create table tmptable(key string, value string, hr string, ds string);
 
 explain extended 
@@ -11,4 +11,4 @@ insert overwrite table tmptable
 select a.* from srcpart a where rand(1) < 0.1 and a.ds = '2008-04-08';
 
 select * from tmptable x sort by x.key,x.value,x.ds,x.hr;
-drop table tmptable;
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_bigdata.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_bigdata.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_bigdata.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_bigdata.q Thu Jul 22 23:57:29 2010
@@ -1,7 +1,7 @@
 set hive.map.aggr.hash.percentmemory = 0.3;
 set hive.mapred.local.mem = 256;
 
-DROP TABLE columnTable_Bigdata;
+
 CREATE table columnTable_Bigdata (key STRING, value STRING)
 ROW FORMAT SERDE
   'org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe'
@@ -15,4 +15,4 @@ INSERT OVERWRITE TABLE columnTable_Bigda
 describe columnTable_Bigdata;
 select count(columnTable_Bigdata.key) from columnTable_Bigdata;
 
-DROP TABLE columnTable_Bigdata;
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_columnar.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_columnar.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_columnar.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_columnar.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-DROP TABLE columnTable;
+
 CREATE table columnTable (key STRING, value STRING)
 ROW FORMAT SERDE
   'org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe'
@@ -12,4 +12,4 @@ describe columnTable;
 
 SELECT columnTable.* FROM columnTable;
 
-DROP TABLE columnTable;
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_default_format.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_default_format.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_default_format.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_default_format.q Thu Jul 22 23:57:29 2010
@@ -14,7 +14,6 @@ SET hive.default.fileformat = TextFile;
 CREATE TABLE textfile_default_format_ctas AS SELECT key,value FROM rcfile_default_format_ctas;
 DESCRIBE EXTENDED textfile_default_format_ctas;
 
-DROP TABLE  rcfile_default_format;
-DROP TABLE  rcfile_default_format_ctas;
-DROP TABLE rcfile_default_format_txtfile;
-DROP TABLE textfile_default_format_ctas;
\ No newline at end of file
+
+
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_lazydecompress.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_lazydecompress.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_lazydecompress.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_lazydecompress.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-DROP TABLE rcfileTableLazyDecompress;
+
 CREATE table rcfileTableLazyDecompress (key STRING, value STRING) STORED AS RCFile;
 
 FROM src
@@ -25,4 +25,3 @@ SELECT key, count(1) FROM rcfileTableLaz
 set mapred.output.compress=false;
 set hive.exec.compress.output=false;
 
-DROP TABLE rcfileTableLazyDecompress;
\ No newline at end of file

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_null_value.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_null_value.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_null_value.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_null_value.q Thu Jul 22 23:57:29 2010
@@ -1,7 +1,7 @@
 CREATE TABLE src1_rc(key STRING, value STRING) STORED AS RCFILE;
 INSERT OVERWRITE TABLE src1_rc SELECT * FROM src1;
 SELECT * FROM src1_rc;
-DROP TABLE src1_rc;
+
 
 CREATE TABLE dest1_rc(c1 INT, c2 STRING, c3 INT, c4 STRING) STORED AS RCFILE;
 
@@ -36,5 +36,5 @@ INSERT OVERWRITE TABLE dest1_rc SELECT c
 
 SELECT dest1_rc.* FROM dest1_rc;
 
-DROP TABLE dest1_rc;
+
 

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_union.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_union.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_union.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/rcfile_union.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-DROP TABLE rcfile_unionTable;
+
 CREATE table rcfile_unionTable (b STRING, c STRING)
 ROW FORMAT SERDE
   'org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe'
@@ -12,4 +12,3 @@ SELECT b AS cola FROM rcfile_unionTable
 UNION ALL
 SELECT c AS cola FROM rcfile_unionTable) s;
 
-DROP TABLE rcfile_unionTable;
\ No newline at end of file

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/reduce_deduplicate.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/reduce_deduplicate.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/reduce_deduplicate.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/reduce_deduplicate.q Thu Jul 22 23:57:29 2010
@@ -2,7 +2,7 @@ set hive.enforce.bucketing = true;
 set hive.exec.reducers.max = 1;
 set hive.exec.script.trust = true;
 
-drop table bucket5_1;
+
 
 CREATE TABLE bucket5_1(key string, value string) CLUSTERED BY (key) INTO 2 BUCKETS;
 explain extended
@@ -15,10 +15,10 @@ select * from src cluster by key;
 select sum(hash(key)),sum(hash(value)) from bucket5_1;
 select sum(hash(key)),sum(hash(value)) from src;
 
-drop table complex_tbl_1;
+
 create table complex_tbl_1(aid string, bid string, t int, ctime string, etime bigint, l string, et string) partitioned by (ds string);
 
-drop table complex_tbl_2;
+
 create table complex_tbl_2(aet string, aes string) partitioned by (ds string);
 
 explain extended
@@ -37,7 +37,7 @@ select s2.* from
 )s
 )s2;
 
-drop table complex_tbl_2;
-drop table complex_tbl_1;
-drop table bucket5_1;
+
+
+
 

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/rename_column.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/rename_column.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/rename_column.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/rename_column.q Thu Jul 22 23:57:29 2010
@@ -22,5 +22,5 @@ DESCRIBE kv_rename_test;
 ALTER TABLE kv_rename_test CHANGE COLUMN a2 a INT AFTER b;
 DESCRIBE kv_rename_test;
 
-DROP TABLE kv_rename_test;
+
 

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/repair.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/repair.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/repair.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/repair.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-DROP TABLE repairtable;
+
 
 CREATE TABLE repairtable(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
 
@@ -13,4 +13,4 @@ MSCK REPAIR TABLE repairtable;
 
 MSCK TABLE repairtable;
 
-DROP TABLE repairtable;
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/sample10.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/sample10.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/sample10.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/sample10.q Thu Jul 22 23:57:29 2010
@@ -21,4 +21,4 @@ select ds, count(1) from srcpartbucket t
 
 select * from srcpartbucket where ds is not null;
 
-drop table srcpartbucket;
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/sample6.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/sample6.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/sample6.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/sample6.q Thu Jul 22 23:57:29 2010
@@ -46,5 +46,5 @@ ORDER BY key, value;
 SELECT s.* FROM empty_bucket TABLESAMPLE (BUCKET 1 OUT OF 2 on key) s
 ORDER BY key, value;
 
-drop table empty_bucket;
-drop table dest1;
+
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/semijoin.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/semijoin.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/semijoin.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/semijoin.q Thu Jul 22 23:57:29 2010
@@ -1,7 +1,7 @@
-drop table t1;
-drop table t2;
-drop table t3;
-drop table t4;
+
+
+
+
 
 create table t1 as select cast(key as int) key, value from src where key <= 10;
 
@@ -77,7 +77,7 @@ select a.key from t3 a left semi join t1
 explain select a.key from t3 a left semi join t2 b on a.key = b.key left outer join t1 c on a.value = c.value sort by a.key;
 select a.key from t3 a left semi join t2 b on a.key = b.key left outer join t1 c on a.value = c.value sort by a.key;
 
-drop table t1;
-drop table t2;
-drop table t3;
-drop table t4;
+
+
+
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/show_tables.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/show_tables.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/show_tables.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/show_tables.q Thu Jul 22 23:57:29 2010
@@ -11,5 +11,5 @@ SHOW TABLES 'shtb_test1|shtb_test2';
 
 SHOW TABLES 'shtb_test1|shtb_test2';
 
-DROP TABLE shtb_test1;
-DROP TABLE shtb_test2;
+
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/skewjoin.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/skewjoin.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/skewjoin.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/skewjoin.q Thu Jul 22 23:57:29 2010
@@ -1,11 +1,11 @@
 set hive.optimize.skewjoin = true;
 set hive.skewjoin.key = 2;
 
-DROP TABLE T1;
-DROP TABLE T2;
-DROP TABLE T3;
-DROP TABLE T4;
-DROP TABLE dest_j1;
+
+
+
+
+
 
 CREATE TABLE T1(key STRING, val STRING) STORED AS TEXTFILE;
 CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;
@@ -131,8 +131,7 @@ SELECT sum(hash(src1.key)), sum(hash(src
 select /*+ mapjoin(v)*/ sum(hash(k.key)), sum(hash(v.val)) from T1 k left outer join T1 v on k.key+1=v.key;
 
 
-DROP TABLE dest_j1;
-DROP TABLE T1;
-DROP TABLE T2;
-DROP TABLE T3;
-DROP TABLE T4;
\ No newline at end of file
+
+
+
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_1.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_1.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_1.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_1.q Thu Jul 22 23:57:29 2010
@@ -1,6 +1,6 @@
-drop table smb_bucket_3;
-drop table smb_bucket_2;
-drop table smb_bucket_1;
+
+
+
 
 create table smb_bucket_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE; 
 create table smb_bucket_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE; 
@@ -47,6 +47,6 @@ select /*+mapjoin(b)*/ * from smb_bucket
 select /*+mapjoin(b)*/ * from smb_bucket_1 a full outer join smb_bucket_2 b on a.key = b.key;
 
  
-drop table smb_bucket_3;
-drop table smb_bucket_2;
-drop table smb_bucket_1;
+
+
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_2.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_2.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_2.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_2.q Thu Jul 22 23:57:29 2010
@@ -1,6 +1,6 @@
-drop table smb_bucket_3;
-drop table smb_bucket_2;
-drop table smb_bucket_1;
+
+
+
 
 create table smb_bucket_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE; 
 create table smb_bucket_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE; 
@@ -47,6 +47,6 @@ select /*+mapjoin(b)*/ * from smb_bucket
 select /*+mapjoin(b)*/ * from smb_bucket_1 a full outer join smb_bucket_3 b on a.key = b.key;
 
  
-drop table smb_bucket_3;
-drop table smb_bucket_2;
-drop table smb_bucket_1;
+
+
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_3.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_3.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_3.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_3.q Thu Jul 22 23:57:29 2010
@@ -1,6 +1,6 @@
-drop table smb_bucket_3;
-drop table smb_bucket_2;
-drop table smb_bucket_1;
+
+
+
 
 create table smb_bucket_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE; 
 create table smb_bucket_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE; 
@@ -47,6 +47,6 @@ select /*+mapjoin(b)*/ * from smb_bucket
 select /*+mapjoin(b)*/ * from smb_bucket_2 a full outer join smb_bucket_3 b on a.key = b.key;
 
  
-drop table smb_bucket_3;
-drop table smb_bucket_2;
-drop table smb_bucket_1;
+
+
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_4.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_4.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_4.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_4.q Thu Jul 22 23:57:29 2010
@@ -1,6 +1,6 @@
-drop table smb_bucket_3;
-drop table smb_bucket_2;
-drop table smb_bucket_1;
+
+
+
 
 create table smb_bucket_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE; 
 create table smb_bucket_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE; 
@@ -66,6 +66,6 @@ select /*+mapjoin(a,b)*/ * from smb_buck
 select /*+mapjoin(a,b)*/ * from smb_bucket_1 a full outer join smb_bucket_2 b on a.key = b.key full outer join smb_bucket_3 c on b.key=c.key;
 
  
-drop table smb_bucket_3;
-drop table smb_bucket_2;
-drop table smb_bucket_1;
+
+
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_5.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_5.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_5.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_5.q Thu Jul 22 23:57:29 2010
@@ -1,6 +1,6 @@
-drop table smb_bucket_3;
-drop table smb_bucket_2;
-drop table smb_bucket_1;
+
+
+
 
 create table smb_bucket_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE; 
 create table smb_bucket_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE; 
@@ -66,6 +66,6 @@ select /*+mapjoin(a,c)*/ * from smb_buck
 select /*+mapjoin(a,c)*/ * from smb_bucket_1 a full outer join smb_bucket_2 b on a.key = b.key full outer join smb_bucket_3 c on b.key=c.key;
 
  
-drop table smb_bucket_3;
-drop table smb_bucket_2;
-drop table smb_bucket_1;
+
+
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_6.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_6.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_6.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_6.q Thu Jul 22 23:57:29 2010
@@ -2,10 +2,10 @@ set hive.enforce.bucketing = true;
 set hive.enforce.sorting = true;
 set hive.exec.reducers.max = 1;
 
-drop table smb_bucket4_1;
+
 CREATE TABLE smb_bucket4_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS RCFILE;
 
-drop table smb_bucket4_2;
+
 CREATE TABLE smb_bucket4_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS RCFILE;
 
 create table smb_join_results(k1 int, v1 string, k2 int, v2 string);
@@ -71,7 +71,7 @@ explain
 select /*+mapjoin(b,c)*/ * from smb_bucket4_1 a join smb_bucket4_2 b on a.key = b.key join smb_bucket4_2 c on b.key = c.key where a.key>1000;
 select /*+mapjoin(b,c)*/ * from smb_bucket4_1 a join smb_bucket4_2 b on a.key = b.key join smb_bucket4_2 c on b.key = c.key where a.key>1000;
 
-drop table smb_join_results;
-drop table normal_join_results;
-drop table smb_bucket4_1;
-drop table smb_bucket4_2;
+
+
+
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_7.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_7.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_7.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_7.q Thu Jul 22 23:57:29 2010
@@ -2,15 +2,15 @@ set hive.enforce.bucketing = true;
 set hive.enforce.sorting = true;
 set hive.exec.reducers.max = 1;
 
-drop table smb_bucket4_1;
+
 CREATE TABLE smb_bucket4_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS;
 
-drop table smb_bucket4_2;
+
 CREATE TABLE smb_bucket4_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS;
 
-drop table smb_join_results;
-drop table smb_join_results_empty_bigtable;
-drop table normal_join_results;
+
+
+
 create table smb_join_results(k1 int, v1 string, k2 int, v2 string);
 create table smb_join_results_empty_bigtable(k1 int, v1 string, k2 int, v2 string);
 create table normal_join_results(k1 int, v1 string, k2 int, v2 string);
@@ -47,8 +47,8 @@ select sum(hash(k1)) as k1, sum(hash(k2)
 select sum(hash(k1)) as k1, sum(hash(k2)) as k2, sum(hash(v1)) as v1, sum(hash(v2)) as v2 from smb_join_results;
 select sum(hash(k1)) as k1, sum(hash(k2)) as k2, sum(hash(v1)) as v1, sum(hash(v2)) as v2 from smb_join_results_empty_bigtable;
 
-drop table smb_join_results;
-drop table smb_join_results_empty_bigtable;
-drop table normal_join_results;
-drop table smb_bucket4_1;
-drop table smb_bucket4_2;
+
+
+
+
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_8.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_8.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_8.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/smb_mapjoin_8.q Thu Jul 22 23:57:29 2010
@@ -1,7 +1,7 @@
 set hive.enforce.bucketing = true;
 set hive.exec.reducers.max = 1;
 
-drop table smb_bucket_input;
+
 create table smb_bucket_input (key int, value string) stored as rcfile;
 load data local inpath '../data/files/smb_bucket_input.rc' into table smb_bucket_input;
 
@@ -10,11 +10,11 @@ set hive.optimize.bucketmapjoin.sortedme
 set hive.enforce.sorting = true;
 set hive.exec.reducers.max = 1;
 
-drop table smb_bucket4_1;
+
 CREATE TABLE smb_bucket4_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS;
-drop table smb_bucket4_2;
+
 CREATE TABLE smb_bucket4_2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS;
-drop table smb_bucket4_3;
+
 CREATE TABLE smb_bucket4_3(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS;
 
 insert overwrite table smb_bucket4_1 select * from smb_bucket_input where key=4 or key=2000 or key=4000;
@@ -84,7 +84,6 @@ insert overwrite table smb_bucket4_3 sel
 select /*+mapjoin(b,c)*/ * from smb_bucket4_1 a full outer join smb_bucket4_2 b on a.key = b.key
 full outer join smb_bucket4_3 c on a.key=c.key;
 
-drop table smb_bucket4_1;
-drop table smb_bucket4_2;
-drop table smb_bucket4_3;
-drop table smb_bucket_input;
\ No newline at end of file
+
+
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/symlink_text_input_format.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/symlink_text_input_format.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/symlink_text_input_format.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/symlink_text_input_format.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-DROP TABLE symlink_text_input_format;
+
 
 EXPLAIN
 CREATE TABLE symlink_text_input_format (key STRING, value STRING) STORED AS INPUTFORMAT 'org.apache.hadoop.hive.ql.io.SymlinkTextInputFormat' OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat';
@@ -20,5 +20,5 @@ EXPLAIN SELECT count(1) FROM symlink_tex
 
 SELECT count(1) FROM symlink_text_input_format;
 
-DROP TABLE symlink_text_input_format;
+
 

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/tablename_with_select.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/tablename_with_select.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/tablename_with_select.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/tablename_with_select.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-DROP TABLE tmp_select;
+
 CREATE TABLE tmp_select(a INT, b STRING);
 DESCRIBE tmp_select;
 
@@ -6,4 +6,4 @@ INSERT OVERWRITE TABLE tmp_select SELECT
 
 SELECT a, b FROM tmp_select ORDER BY a;
 
-DROP TABLE tmp_select;
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/transform1.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/transform1.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/transform1.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/transform1.q Thu Jul 22 23:57:29 2010
@@ -1,4 +1,4 @@
-drop table transform1_t1;
+
 create table transform1_t1(a string, b string);
 
 EXPLAIN
@@ -6,9 +6,9 @@ SELECT transform(*) USING 'cat' AS (col 
 
 SELECT transform(*) USING 'cat' AS (col array<bigint>) FROM transform1_t1;
 
-drop table transform1_t1;
 
-drop table transform1_t2;
+
+
 create table transform1_t2(col array<int>);
 
 insert overwrite table transform1_t2
@@ -19,5 +19,5 @@ SELECT transform('0\0021\0022') USING 'c
 
 SELECT transform('0\0021\0022') USING 'cat' AS (col array<int>) FROM transform1_t2;
 
-drop table transform1_t2;
+
 

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_concat_insert1.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_concat_insert1.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_concat_insert1.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_concat_insert1.q Thu Jul 22 23:57:29 2010
@@ -5,4 +5,3 @@ INSERT OVERWRITE TABLE dest1 SELECT '123
 
 SELECT dest1.* FROM dest1;
 
-DROP TABLE dest1;
\ No newline at end of file

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_concat_insert2.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_concat_insert2.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_concat_insert2.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_concat_insert2.q Thu Jul 22 23:57:29 2010
@@ -5,4 +5,4 @@ INSERT OVERWRITE TABLE dest1 SELECT conc
 
 SELECT dest1.* FROM dest1;
 
-DROP TABLE dest1;
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_field.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_field.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_field.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_field.q Thu Jul 22 23:57:29 2010
@@ -23,7 +23,7 @@ SELECT
   field(4, 1, 2, NULL, 4)
 FROM src LIMIT 1;
 
-DROP TABLE test_table;
+
 CREATE TABLE test_table(col1 STRING, col2 STRING) STORED AS TEXTFILE;
 LOAD DATA LOCAL INPATH '../data/files/kv1.txt' INTO TABLE test_table;
 
@@ -40,7 +40,7 @@ select col1,col2,
   field(col1, col2, "66")
 from test_table where col1="86" or col1="66";
 
-DROP TABLE test_table1;
+
 CREATE TABLE test_table1(col1 int, col2 string) STORED AS TEXTFILE;
 LOAD DATA LOCAL INPATH '../data/files/kv1.txt' INTO TABLE test_table1;
 

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_length.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_length.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_length.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_length.q Thu Jul 22 23:57:29 2010
@@ -12,4 +12,3 @@ CREATE TABLE dest1(name STRING) STORED A
 LOAD DATA LOCAL INPATH '../data/files/kv4.txt' INTO TABLE dest1;
 EXPLAIN SELECT length(dest1.name) FROM dest1;
 SELECT length(dest1.name) FROM dest1;
-DROP TABLE dest1;

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_reverse.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_reverse.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_reverse.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_reverse.q Thu Jul 22 23:57:29 2010
@@ -13,4 +13,3 @@ DROP TABLE dest1;
 CREATE TABLE dest1(name STRING) STORED AS TEXTFILE;
 LOAD DATA LOCAL INPATH '../data/files/kv4.txt' INTO TABLE dest1;
 SELECT count(1) FROM dest1 WHERE reverse(dest1.name) = _UTF-8 0xE993AEE982B5;
-DROP TABLE dest1;

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_sentences.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_sentences.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_sentences.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/udf_sentences.q Thu Jul 22 23:57:29 2010
@@ -5,10 +5,10 @@ SELECT explode(sentences(unhex("48697665
 INSERT OVERWRITE TABLE sent_tmp2
 SELECT explode(val) AS val FROM sent_tmp;
 SELECT hex(val) FROM sent_tmp2;
+
 DROP TABLE sent_tmp;
 DROP TABLE sent_tmp2;
 
-
 CREATE TABLE sent_tmp (val array<string>);
 CREATE TABLE sent_tmp2 (val string);
 INSERT OVERWRITE TABLE sent_tmp
@@ -16,7 +16,7 @@ SELECT explode(sentences(unhex("48697665
 INSERT OVERWRITE TABLE sent_tmp2
 SELECT explode(val) AS val FROM sent_tmp;
 SELECT hex(val) FROM sent_tmp2;
-DROP TABLE sent_tmp;
-DROP TABLE sent_tmp2;
+
+
 
 SELECT sentences("Hive is an excellent tool for data querying\; and perhaps more versatile than machine translation!! Multiple, ill-formed sentences...confounding punctuation--and yet this UDF still works!!!!") FROM src LIMIT 1;

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/union10.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/union10.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/union10.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/union10.q Thu Jul 22 23:57:29 2010
@@ -2,7 +2,6 @@ set hive.map.aggr = true;
 
 -- union case: all subqueries are a map-reduce jobs, 3 way union, same input for all sub-queries, followed by filesink
 
-drop table tmptable;
 create table tmptable(key string, value int);
 
 explain 
@@ -24,5 +23,3 @@ insert overwrite table tmptable
 
 select * from tmptable x sort by x.key;
 
-drop table tmptable;
-

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/union12.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/union12.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/union12.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/union12.q Thu Jul 22 23:57:29 2010
@@ -2,7 +2,6 @@ set hive.map.aggr = true;
 
 -- union case: all subqueries are a map-reduce jobs, 3 way union, different inputs for all sub-queries, followed by filesink
 
-drop table tmptable;
 create table tmptable(key string, value int);
 
 explain 
@@ -22,5 +21,3 @@ insert overwrite table tmptable
                                             select 'tst3' as key, count(1) as value from srcbucket s3) unionsrc;
 
 select * from tmptable x sort by x.key;
-
-drop table tmptable;

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/union17.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/union17.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/union17.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/union17.q Thu Jul 22 23:57:29 2010
@@ -1,6 +1,3 @@
-drop table DEST1;
-drop table DEST2;
-
 CREATE TABLE DEST1(key STRING, value STRING) STORED AS TEXTFILE;
 CREATE TABLE DEST2(key STRING, val1 STRING, val2 STRING) STORED AS TEXTFILE;
 
@@ -21,6 +18,3 @@ INSERT OVERWRITE TABLE DEST2 SELECT unio
 
 SELECT DEST1.* FROM DEST1;
 SELECT DEST2.* FROM DEST2;
-
-drop table DEST1;
-drop table DEST2;

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/union18.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/union18.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/union18.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/union18.q Thu Jul 22 23:57:29 2010
@@ -1,6 +1,3 @@
-drop table DEST1;
-drop table DEST2;
-
 CREATE TABLE DEST1(key STRING, value STRING) STORED AS TEXTFILE;
 CREATE TABLE DEST2(key STRING, val1 STRING, val2 STRING) STORED AS TEXTFILE;
 
@@ -21,6 +18,3 @@ INSERT OVERWRITE TABLE DEST2 SELECT unio
 
 SELECT DEST1.* FROM DEST1 SORT BY DEST1.key, DEST1.value;
 SELECT DEST2.* FROM DEST2 SORT BY DEST2.key, DEST2.val1, DEST2.val2;
-
-drop table DEST1;
-drop table DEST2;

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/union19.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/union19.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/union19.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/union19.q Thu Jul 22 23:57:29 2010
@@ -1,5 +1,5 @@
-drop table DEST1;
-drop table DEST2;
+
+
 
 CREATE TABLE DEST1(key STRING, value STRING) STORED AS TEXTFILE;
 CREATE TABLE DEST2(key STRING, val1 STRING, val2 STRING) STORED AS TEXTFILE;
@@ -22,5 +22,5 @@ INSERT OVERWRITE TABLE DEST2 SELECT unio
 SELECT DEST1.* FROM DEST1 SORT BY DEST1.key, DEST1.value;
 SELECT DEST2.* FROM DEST2 SORT BY DEST2.key, DEST2.val1, DEST2.val2;
 
-drop table DEST1;
-drop table DEST2;
+
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/union22.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/union22.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/union22.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/union22.q Thu Jul 22 23:57:29 2010
@@ -1,7 +1,7 @@
-drop table dst_union22;
+
 create table dst_union22(k1 string, k2 string, k3 string, k4 string) partitioned by (ds string);
 
-drop table dst_union22_delta;
+
 create table dst_union22_delta(k0 string, k1 string, k2 string, k3 string, k4 string, k5 string) partitioned by (ds string);
 
 insert overwrite table dst_union22 partition (ds='1')
@@ -40,5 +40,5 @@ subq;
 
 select * from dst_union22 where ds = '2' order by k1;
 
-drop table dst_union22;
-drop table dst_union22_delta;
+
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/union3.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/union3.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/union3.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/union3.q Thu Jul 22 23:57:29 2010
@@ -17,7 +17,7 @@ FROM (
   FROM (SELECT * FROM src LIMIT 1) s2
 ) a;
 
-DROP TABLE union_out;
+
 
 CREATE TABLE union_out (id int);
 

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/union4.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/union4.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/union4.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/union4.q Thu Jul 22 23:57:29 2010
@@ -2,7 +2,7 @@ set hive.map.aggr = true;
 
 -- union case: both subqueries are map-reduce jobs on same input, followed by filesink
 
-drop table tmptable;
+
 create table tmptable(key string, value int);
 
 explain 
@@ -18,4 +18,4 @@ select unionsrc.key, unionsrc.value FROM
 
 select * from tmptable x sort by x.key;
 
-drop table tmptable;
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/union6.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/union6.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/union6.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/union6.q Thu Jul 22 23:57:29 2010
@@ -2,7 +2,7 @@ set hive.map.aggr = true;
 
 -- union case: 1 subquery is a map-reduce job, different inputs for sub-queries, followed by filesink
 
-drop table tmptable;
+
 create table tmptable(key string, value string);
 
 explain 
@@ -18,4 +18,4 @@ select unionsrc.key, unionsrc.value FROM
 
 select * from tmptable x sort by x.key, x.value;
 
-drop table tmptable;
+

Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/uniquejoin.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/uniquejoin.q?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/uniquejoin.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/uniquejoin.q Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-DROP TABLE T1;
-DROP TABLE T2;
-DROP TABLE T3;
-
 CREATE TABLE T1(key STRING, val STRING) STORED AS TEXTFILE;
 CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;
 CREATE TABLE T3(key STRING, val STRING) STORED AS TEXTFILE;
@@ -27,7 +23,3 @@ SELECT a.key, b.key, c.key;
 
 FROM UNIQUEJOIN PRESERVE T1 a (a.key), T2 b(b.key)
 SELECT a.key, b.key;
-
-DROP TABLE T1;
-DROP TABLE T2;
-DROP TABLE T3;

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/addpart1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/addpart1.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/addpart1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/addpart1.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table addpart1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table addpart1
-POSTHOOK: type: DROPTABLE
 PREHOOK: query: create table addpart1 (a int) partitioned by (b string, c string)
 PREHOOK: type: CREATETABLE
 POSTHOOK: query: create table addpart1 (a int) partitioned by (b string, c string)

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/alter_non_native.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/alter_non_native.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/alter_non_native.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/alter_non_native.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE non_native1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE non_native1
-POSTHOOK: type: DROPTABLE
 PREHOOK: query: CREATE TABLE non_native1(key int, value string) 
 STORED BY 'org.apache.hadoop.hive.ql.metadata.DefaultStorageHandler'
 PREHOOK: type: CREATETABLE

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/altern1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/altern1.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/altern1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/altern1.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table altern1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table altern1
-POSTHOOK: type: DROPTABLE
 PREHOOK: query: create table altern1(a int, b int) partitioned by (ds string)
 PREHOOK: type: CREATETABLE
 POSTHOOK: query: create table altern1(a int, b int) partitioned by (ds string)

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/create_insert_outputformat.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/create_insert_outputformat.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/create_insert_outputformat.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/create_insert_outputformat.q.out Thu Jul 22 23:57:29 2010
@@ -1,5 +1 @@
-PREHOOK: query: DROP TABLE table_test_output_fomat
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE table_test_output_fomat
-POSTHOOK: type: DROPTABLE
 FAILED: Error in semantic analysis: Output Format must implement HiveOutputFormat, otherwise it should be either IgnoreKeyTextOutputFormat or SequenceFileOutputFormat

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/create_view_failure1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/create_view_failure1.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/create_view_failure1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/create_view_failure1.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE xxx12
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE xxx12
-POSTHOOK: type: DROPTABLE
 PREHOOK: query: DROP VIEW xxx12
 PREHOOK: type: DROPVIEW
 POSTHOOK: query: DROP VIEW xxx12
@@ -15,6 +11,6 @@ POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@xxx12
 PREHOOK: query: CREATE VIEW xxx12 AS SELECT key FROM src
 PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/Users/jsichi/open/hive-trunk/build/ql/tmp/194996627/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-43-17_176_344624644454180304/10000
 FAILED: Error in metadata: AlreadyExistsException(message:Table xxx12 already exists)
 FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/create_view_failure2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/create_view_failure2.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/create_view_failure2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/create_view_failure2.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE xxx4
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE xxx4
-POSTHOOK: type: DROPTABLE
 PREHOOK: query: DROP VIEW xxx4
 PREHOOK: type: DROPVIEW
 POSTHOOK: query: DROP VIEW xxx4
@@ -9,11 +5,11 @@ POSTHOOK: type: DROPVIEW
 PREHOOK: query: -- views and tables share the same namespace
 CREATE VIEW xxx4 AS SELECT key FROM src
 PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/Users/jsichi/open/hive-trunk/build/ql/tmp/1327466632/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-43-17_460_630133670553073559/10000
 POSTHOOK: query: -- views and tables share the same namespace
 CREATE VIEW xxx4 AS SELECT key FROM src
 POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/Users/jsichi/open/hive-trunk/build/ql/tmp/1327466632/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-43-17_460_630133670553073559/10000
 POSTHOOK: Output: default@xxx4
 PREHOOK: query: CREATE TABLE xxx4(key int)
 PREHOOK: type: CREATETABLE

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/ctas.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/ctas.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/ctas.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/ctas.q.out Thu Jul 22 23:57:29 2010
@@ -1,5 +1 @@
-PREHOOK: query: drop table nzhang_ctas4
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table nzhang_ctas4
-POSTHOOK: type: DROPTABLE
 FAILED: Error in semantic analysis: CREATE-TABLE-AS-SELECT cannot create external table.

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/ddltime.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/ddltime.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/ddltime.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/ddltime.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table T2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table T2
-POSTHOOK: type: DROPTABLE
 PREHOOK: query: create table T2 like srcpart
 PREHOOK: type: CREATETABLE
 POSTHOOK: query: create table T2 like srcpart

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/deletejar.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/deletejar.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/deletejar.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/deletejar.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE DELETEJAR
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE DELETEJAR
-POSTHOOK: type: DROPTABLE
 PREHOOK: query: CREATE TABLE DELETEJAR(KEY STRING, VALUE STRING) ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.TestSerDe' STORED AS TEXTFILE
 PREHOOK: type: CREATETABLE
 FAILED: Error in metadata: Cannot validate serde: org.apache.hadoop.hive.serde2.TestSerDe

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/drop_view_failure1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/drop_view_failure1.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/drop_view_failure1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/drop_view_failure1.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE xxx1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE xxx1
-POSTHOOK: type: DROPTABLE
 PREHOOK: query: CREATE TABLE xxx1(key int)
 PREHOOK: type: CREATETABLE
 POSTHOOK: query: CREATE TABLE xxx1(key int)

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/dyn_part1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/dyn_part1.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/dyn_part1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/dyn_part1.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table dynamic_partition
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table dynamic_partition
-POSTHOOK: type: DROPTABLE
 PREHOOK: query: create table dynamic_partition (key string) partitioned by (value string)
 PREHOOK: type: CREATETABLE
 POSTHOOK: query: create table dynamic_partition (key string) partitioned by (value string)

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/dyn_part2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/dyn_part2.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/dyn_part2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/dyn_part2.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table nzhang_part1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table nzhang_part1
-POSTHOOK: type: DROPTABLE
 PREHOOK: query: create table nzhang_part1 (key string, value string) partitioned by (ds string, hr string)
 PREHOOK: type: CREATETABLE
 POSTHOOK: query: create table nzhang_part1 (key string, value string) partitioned by (ds string, hr string)

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/external1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/external1.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/external1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/external1.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table external1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table external1
-POSTHOOK: type: DROPTABLE
 PREHOOK: query: create external table external1(a int, b int) location 'invalidscheme://data.s3ndemo.hive/kv'
 PREHOOK: type: CREATETABLE
 FAILED: Error in metadata: MetaException(message:Got exception: java.io.IOException No FileSystem for scheme: invalidscheme)

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/external2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/external2.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/external2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/external2.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table external2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table external2
-POSTHOOK: type: DROPTABLE
 PREHOOK: query: create external table external2(a int, b int) partitioned by (ds string)
 PREHOOK: type: CREATETABLE
 POSTHOOK: query: create external table external2(a int, b int) partitioned by (ds string)

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_create_tbl1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_create_tbl1.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_create_tbl1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_create_tbl1.q.out Thu Jul 22 23:57:29 2010
@@ -1,5 +1 @@
-PREHOOK: query: DROP TABLE inv_valid_tbl1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE inv_valid_tbl1
-POSTHOOK: type: DROPTABLE
 FAILED: Error in semantic analysis: DATE, DATETIME, and TIMESTAMP types aren't supported yet. Please use STRING instead.

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/invalidate_view1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/invalidate_view1.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/invalidate_view1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/invalidate_view1.q.out Thu Jul 22 23:57:29 2010
@@ -6,10 +6,6 @@ PREHOOK: query: DROP VIEW xxx9
 PREHOOK: type: DROPVIEW
 POSTHOOK: query: DROP VIEW xxx9
 POSTHOOK: type: DROPVIEW
-PREHOOK: query: DROP TABLE xxx10
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE xxx10
-POSTHOOK: type: DROPTABLE
 PREHOOK: query: -- create two levels of view reference, then invalidate intermediate view
 -- by dropping a column from underlying table, and verify that
 -- querying outermost view results in full error context
@@ -23,17 +19,17 @@ POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@xxx10
 PREHOOK: query: CREATE VIEW xxx9 AS SELECT * FROM xxx10
 PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/Users/jsichi/open/hive-trunk/build/ql/tmp/1720348561/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-43-31_670_5934534815948379913/10000
 POSTHOOK: query: CREATE VIEW xxx9 AS SELECT * FROM xxx10
 POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/Users/jsichi/open/hive-trunk/build/ql/tmp/1720348561/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-43-31_670_5934534815948379913/10000
 POSTHOOK: Output: default@xxx9
 PREHOOK: query: CREATE VIEW xxx8 AS SELECT * FROM xxx9 xxx
 PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/Users/jsichi/open/hive-trunk/build/ql/tmp/1891844844/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-43-31_706_855643045189903710/10000
 POSTHOOK: query: CREATE VIEW xxx8 AS SELECT * FROM xxx9 xxx
 POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/Users/jsichi/open/hive-trunk/build/ql/tmp/1891844844/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-43-31_706_855643045189903710/10000
 POSTHOOK: Output: default@xxx8
 PREHOOK: query: ALTER TABLE xxx10 REPLACE COLUMNS (key int)
 PREHOOK: type: ALTERTABLE_REPLACECOLS

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/load_non_native.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/load_non_native.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/load_non_native.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/load_non_native.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE non_native2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE non_native2
-POSTHOOK: type: DROPTABLE
 PREHOOK: query: CREATE TABLE non_native2(key int, value string) 
 STORED BY 'org.apache.hadoop.hive.ql.metadata.DefaultStorageHandler'
 PREHOOK: type: CREATETABLE

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat.q.out Thu Jul 22 23:57:29 2010
@@ -1,16 +1,14 @@
 PREHOOK: query: -- test for loading into tables with the correct file format
 -- test for loading into partitions with the correct file format
 
-DROP TABLE load_wrong_fileformat_T1
-PREHOOK: type: DROPTABLE
+
+CREATE TABLE load_wrong_fileformat_T1(name STRING) STORED AS SEQUENCEFILE
+PREHOOK: type: CREATETABLE
 POSTHOOK: query: -- test for loading into tables with the correct file format
 -- test for loading into partitions with the correct file format
 
-DROP TABLE load_wrong_fileformat_T1
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE load_wrong_fileformat_T1(name STRING) STORED AS SEQUENCEFILE
-PREHOOK: type: CREATETABLE
-POSTHOOK: query: CREATE TABLE load_wrong_fileformat_T1(name STRING) STORED AS SEQUENCEFILE
+
+CREATE TABLE load_wrong_fileformat_T1(name STRING) STORED AS SEQUENCEFILE
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@load_wrong_fileformat_T1
 PREHOOK: query: LOAD DATA LOCAL INPATH '../data/files/kv1.txt' INTO TABLE load_wrong_fileformat_T1

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat_rc_seq.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat_rc_seq.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat_rc_seq.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat_rc_seq.q.out Thu Jul 22 23:57:29 2010
@@ -1,16 +1,14 @@
 PREHOOK: query: -- test for loading into tables with the correct file format
 -- test for loading into partitions with the correct file format
 
-DROP TABLE T1
-PREHOOK: type: DROPTABLE
+
+CREATE TABLE T1(name STRING) STORED AS RCFILE
+PREHOOK: type: CREATETABLE
 POSTHOOK: query: -- test for loading into tables with the correct file format
 -- test for loading into partitions with the correct file format
 
-DROP TABLE T1
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE T1(name STRING) STORED AS RCFILE
-PREHOOK: type: CREATETABLE
-POSTHOOK: query: CREATE TABLE T1(name STRING) STORED AS RCFILE
+
+CREATE TABLE T1(name STRING) STORED AS RCFILE
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@T1
 PREHOOK: query: LOAD DATA LOCAL INPATH '../data/files/kv1.seq' INTO TABLE T1

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat_txt_seq.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat_txt_seq.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat_txt_seq.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat_txt_seq.q.out Thu Jul 22 23:57:29 2010
@@ -1,16 +1,14 @@
 PREHOOK: query: -- test for loading into tables with the correct file format
 -- test for loading into partitions with the correct file format
 
-DROP TABLE T1
-PREHOOK: type: DROPTABLE
+
+CREATE TABLE T1(name STRING) STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
 POSTHOOK: query: -- test for loading into tables with the correct file format
 -- test for loading into partitions with the correct file format
 
-DROP TABLE T1
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE T1(name STRING) STORED AS TEXTFILE
-PREHOOK: type: CREATETABLE
-POSTHOOK: query: CREATE TABLE T1(name STRING) STORED AS TEXTFILE
+
+CREATE TABLE T1(name STRING) STORED AS TEXTFILE
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@T1
 PREHOOK: query: LOAD DATA LOCAL INPATH '../data/files/kv1.seq' INTO TABLE T1

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/nopart_insert.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/nopart_insert.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/nopart_insert.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/nopart_insert.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE nopart_insert
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE nopart_insert
-POSTHOOK: type: DROPTABLE
 PREHOOK: query: CREATE TABLE nopart_insert(a STRING, b STRING) PARTITIONED BY (ds STRING)
 PREHOOK: type: CREATETABLE
 POSTHOOK: query: CREATE TABLE nopart_insert(a STRING, b STRING) PARTITIONED BY (ds STRING)

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/nopart_load.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/nopart_load.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/nopart_load.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/nopart_load.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE nopart_load
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE nopart_load
-POSTHOOK: type: DROPTABLE
 PREHOOK: query: CREATE TABLE nopart_load(a STRING, b STRING) PARTITIONED BY (ds STRING)
 PREHOOK: type: CREATETABLE
 POSTHOOK: query: CREATE TABLE nopart_load(a STRING, b STRING) PARTITIONED BY (ds STRING)

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/smb_bucketmapjoin.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/smb_bucketmapjoin.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/smb_bucketmapjoin.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/smb_bucketmapjoin.q.out Thu Jul 22 23:57:29 2010
@@ -1,16 +1,8 @@
-PREHOOK: query: drop table smb_bucket4_1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table smb_bucket4_1
-POSTHOOK: type: DROPTABLE
 PREHOOK: query: CREATE TABLE smb_bucket4_1(key int, value string) CLUSTERED BY (key) INTO 2 BUCKETS
 PREHOOK: type: CREATETABLE
 POSTHOOK: query: CREATE TABLE smb_bucket4_1(key int, value string) CLUSTERED BY (key) INTO 2 BUCKETS
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@smb_bucket4_1
-PREHOOK: query: drop table smb_bucket4_2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table smb_bucket4_2
-POSTHOOK: type: DROPTABLE
 PREHOOK: query: CREATE TABLE smb_bucket4_2(key int, value string) CLUSTERED BY (key) INTO 2 BUCKETS
 PREHOOK: type: CREATETABLE
 POSTHOOK: query: CREATE TABLE smb_bucket4_2(key int, value string) CLUSTERED BY (key) INTO 2 BUCKETS

Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/union2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/union2.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/union2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/union2.q.out Thu Jul 22 23:57:29 2010
@@ -1,11 +1,3 @@
-PREHOOK: query: drop table union2_t1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table union2_t1
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: drop table union2_t2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table union2_t2
-POSTHOOK: type: DROPTABLE
 PREHOOK: query: create table if not exists union2_t1(r string, c string, v string)
 PREHOOK: type: CREATETABLE
 POSTHOOK: query: create table if not exists union2_t1(r string, c string, v string)

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/add_part_exist.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/add_part_exist.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/add_part_exist.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/add_part_exist.q.out Thu Jul 22 23:57:29 2010
@@ -49,8 +49,3 @@ POSTHOOK: type: SHOWPARTITIONS
 ds=2010-01-01
 ds=2010-01-02
 ds=2010-01-03
-PREHOOK: query: DROP TABLE add_part_test
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE add_part_test
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@add_part_test