You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by na...@apache.org on 2013/04/09 08:33:23 UTC

svn commit: r1465903 [3/13] - in /hive/trunk/ql/src/test: queries/clientpositive/ results/clientpositive/

Modified: hive/trunk/ql/src/test/results/clientpositive/binary_output_format.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/binary_output_format.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/binary_output_format.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/binary_output_format.q.out Tue Apr  9 06:33:16 2013
@@ -1,5 +1,4 @@
-PREHOOK: query: -- Create a table with binary output format
-CREATE TABLE dest1(mydata STRING)
+PREHOOK: query: CREATE TABLE dest1(mydata STRING)
 ROW FORMAT SERDE
   'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
 WITH SERDEPROPERTIES (
@@ -9,8 +8,7 @@ STORED AS
   INPUTFORMAT 'org.apache.hadoop.mapred.TextInputFormat'
   OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat'
 PREHOOK: type: CREATETABLE
-POSTHOOK: query: -- Create a table with binary output format
-CREATE TABLE dest1(mydata STRING)
+POSTHOOK: query: CREATE TABLE dest1(mydata STRING)
 ROW FORMAT SERDE
   'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
 WITH SERDEPROPERTIES (
@@ -21,8 +19,7 @@ STORED AS
   OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat'
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@dest1
-PREHOOK: query: -- Insert into that table using transform
-EXPLAIN EXTENDED
+PREHOOK: query: EXPLAIN EXTENDED
 INSERT OVERWRITE TABLE dest1
 SELECT TRANSFORM(*)
   USING 'cat'
@@ -35,8 +32,7 @@ SELECT TRANSFORM(*)
     RECORDREADER 'org.apache.hadoop.hive.ql.exec.BinaryRecordReader'
 FROM src
 PREHOOK: type: QUERY
-POSTHOOK: query: -- Insert into that table using transform
-EXPLAIN EXTENDED
+POSTHOOK: query: EXPLAIN EXTENDED
 INSERT OVERWRITE TABLE dest1
 SELECT TRANSFORM(*)
   USING 'cat'
@@ -372,13 +368,11 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
 POSTHOOK: Output: default@dest1
 POSTHOOK: Lineage: dest1.mydata SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
-PREHOOK: query: -- Test the result
-SELECT * FROM dest1
+PREHOOK: query: SELECT * FROM dest1
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest1
 #### A masked pattern was here ####
-POSTHOOK: query: -- Test the result
-SELECT * FROM dest1
+POSTHOOK: query: SELECT * FROM dest1
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest1
 #### A masked pattern was here ####

Modified: hive/trunk/ql/src/test/results/clientpositive/binary_table_bincolserde.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/binary_table_bincolserde.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/binary_table_bincolserde.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/binary_table_bincolserde.q.out Tue Apr  9 06:33:16 2013
@@ -2,13 +2,9 @@ PREHOOK: query: drop table ba_test
 PREHOOK: type: DROPTABLE
 POSTHOOK: query: drop table ba_test
 POSTHOOK: type: DROPTABLE
-PREHOOK: query: -- Tests everything in binary_table_colserde.q + uses LazyBinaryColumnarSerde
-
-create table ba_test (ba_key binary, ba_val binary) stored as rcfile
+PREHOOK: query: create table ba_test (ba_key binary, ba_val binary) stored as rcfile
 PREHOOK: type: CREATETABLE
-POSTHOOK: query: -- Tests everything in binary_table_colserde.q + uses LazyBinaryColumnarSerde
-
-create table ba_test (ba_key binary, ba_val binary) stored as rcfile
+POSTHOOK: query: create table ba_test (ba_key binary, ba_val binary) stored as rcfile
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@ba_test
 PREHOOK: query: alter table ba_test set serde 'org.apache.hadoop.hive.serde2.columnar.LazyBinaryColumnarSerDe'

Modified: hive/trunk/ql/src/test/results/clientpositive/binary_table_colserde.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/binary_table_colserde.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/binary_table_colserde.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/binary_table_colserde.q.out Tue Apr  9 06:33:16 2013
@@ -2,13 +2,9 @@ PREHOOK: query: drop table ba_test
 PREHOOK: type: DROPTABLE
 POSTHOOK: query: drop table ba_test
 POSTHOOK: type: DROPTABLE
-PREHOOK: query: -- Everything in ba_table1.q + columnar serde in RCFILE.
-
-create table ba_test (ba_key binary, ba_val binary) stored as rcfile
+PREHOOK: query: create table ba_test (ba_key binary, ba_val binary) stored as rcfile
 PREHOOK: type: CREATETABLE
-POSTHOOK: query: -- Everything in ba_table1.q + columnar serde in RCFILE.
-
-create table ba_test (ba_key binary, ba_val binary) stored as rcfile
+POSTHOOK: query: create table ba_test (ba_key binary, ba_val binary) stored as rcfile
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@ba_test
 PREHOOK: query: describe extended ba_test

Modified: hive/trunk/ql/src/test/results/clientpositive/bucket_groupby.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/bucket_groupby.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/bucket_groupby.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/bucket_groupby.q.out Tue Apr  9 06:33:16 2013
@@ -1180,13 +1180,11 @@ STAGE PLANS:
       limit: -1
 
 
-PREHOOK: query: -- number of buckets cannot be changed, so drop the table
-drop table clustergroupby
+PREHOOK: query: drop table clustergroupby
 PREHOOK: type: DROPTABLE
 PREHOOK: Input: default@clustergroupby
 PREHOOK: Output: default@clustergroupby
-POSTHOOK: query: -- number of buckets cannot be changed, so drop the table
-drop table clustergroupby
+POSTHOOK: query: drop table clustergroupby
 POSTHOOK: type: DROPTABLE
 POSTHOOK: Input: default@clustergroupby
 POSTHOOK: Output: default@clustergroupby
@@ -1585,13 +1583,11 @@ POSTHOOK: Lineage: clustergroupby PARTIT
 111	1
 113	2
 114	1
-PREHOOK: query: -- number of buckets cannot be changed, so drop the table
-drop table clustergroupby
+PREHOOK: query: drop table clustergroupby
 PREHOOK: type: DROPTABLE
 PREHOOK: Input: default@clustergroupby
 PREHOOK: Output: default@clustergroupby
-POSTHOOK: query: -- number of buckets cannot be changed, so drop the table
-drop table clustergroupby
+POSTHOOK: query: drop table clustergroupby
 POSTHOOK: type: DROPTABLE
 POSTHOOK: Input: default@clustergroupby
 POSTHOOK: Output: default@clustergroupby

Modified: hive/trunk/ql/src/test/results/clientpositive/bucket_map_join_1.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/bucket_map_join_1.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/bucket_map_join_1.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/bucket_map_join_1.q.out Tue Apr  9 06:33:16 2013
@@ -32,18 +32,10 @@ PREHOOK: Output: default@table2
 POSTHOOK: query: load data local inpath '../data/files/SortCol2Col1.txt' overwrite into table table2
 POSTHOOK: type: LOAD
 POSTHOOK: Output: default@table2
-PREHOOK: query: -- The tables are bucketed in same columns in different order,
--- but sorted in different column orders
--- Neither bucketed map-join, nor sort-merge join should be performed
-
-explain extended
+PREHOOK: query: explain extended
 select /*+ mapjoin(b) */ count(*) from table1 a join table2 b on a.key=b.key and a.value=b.value
 PREHOOK: type: QUERY
-POSTHOOK: query: -- The tables are bucketed in same columns in different order,
--- but sorted in different column orders
--- Neither bucketed map-join, nor sort-merge join should be performed
-
-explain extended
+POSTHOOK: query: explain extended
 select /*+ mapjoin(b) */ count(*) from table1 a join table2 b on a.key=b.key and a.value=b.value
 POSTHOOK: type: QUERY
 ABSTRACT SYNTAX TREE:

Modified: hive/trunk/ql/src/test/results/clientpositive/bucket_map_join_2.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/bucket_map_join_2.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/bucket_map_join_2.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/bucket_map_join_2.q.out Tue Apr  9 06:33:16 2013
@@ -32,18 +32,10 @@ PREHOOK: Output: default@table2
 POSTHOOK: query: load data local inpath '../data/files/SortCol2Col1.txt' overwrite into table table2
 POSTHOOK: type: LOAD
 POSTHOOK: Output: default@table2
-PREHOOK: query: -- The tables are bucketed in same columns in different order,
--- but sorted in different column orders
--- Neither bucketed map-join, nor sort-merge join should be performed
-
-explain extended
+PREHOOK: query: explain extended
 select /*+ mapjoin(b) */ count(*) from table1 a join table2 b on a.key=b.key and a.value=b.value
 PREHOOK: type: QUERY
-POSTHOOK: query: -- The tables are bucketed in same columns in different order,
--- but sorted in different column orders
--- Neither bucketed map-join, nor sort-merge join should be performed
-
-explain extended
+POSTHOOK: query: explain extended
 select /*+ mapjoin(b) */ count(*) from table1 a join table2 b on a.key=b.key and a.value=b.value
 POSTHOOK: type: QUERY
 ABSTRACT SYNTAX TREE:

Modified: hive/trunk/ql/src/test/results/clientpositive/bucketcontext_1.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/bucketcontext_1.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/bucketcontext_1.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/bucketcontext_1.q.out Tue Apr  9 06:33:16 2013
@@ -1,8 +1,6 @@
-PREHOOK: query: -- small 1 part, 2 bucket & big 2 part, 4 bucket
-CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
+PREHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
 PREHOOK: type: CREATETABLE
-POSTHOOK: query: -- small 1 part, 2 bucket & big 2 part, 4 bucket
-CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
+POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@bucket_small
 PREHOOK: query: load data local inpath '../data/files/srcsortbucket1outof4.txt' INTO TABLE bucket_small partition(ds='2008-04-08')

Modified: hive/trunk/ql/src/test/results/clientpositive/bucketcontext_2.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/bucketcontext_2.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/bucketcontext_2.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/bucketcontext_2.q.out Tue Apr  9 06:33:16 2013
@@ -1,8 +1,6 @@
-PREHOOK: query: -- small 1 part, 4 bucket & big 2 part, 2 bucket
-CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE
+PREHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE
 PREHOOK: type: CREATETABLE
-POSTHOOK: query: -- small 1 part, 4 bucket & big 2 part, 2 bucket
-CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE
+POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@bucket_small
 PREHOOK: query: load data local inpath '../data/files/srcsortbucket1outof4.txt' INTO TABLE bucket_small partition(ds='2008-04-08')

Modified: hive/trunk/ql/src/test/results/clientpositive/bucketcontext_3.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/bucketcontext_3.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/bucketcontext_3.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/bucketcontext_3.q.out Tue Apr  9 06:33:16 2013
@@ -1,8 +1,6 @@
-PREHOOK: query: -- small 2 part, 2 bucket & big 1 part, 4 bucket
-CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
+PREHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
 PREHOOK: type: CREATETABLE
-POSTHOOK: query: -- small 2 part, 2 bucket & big 1 part, 4 bucket
-CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
+POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@bucket_small
 PREHOOK: query: load data local inpath '../data/files/srcsortbucket1outof4.txt' INTO TABLE bucket_small partition(ds='2008-04-08')

Modified: hive/trunk/ql/src/test/results/clientpositive/bucketcontext_4.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/bucketcontext_4.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/bucketcontext_4.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/bucketcontext_4.q.out Tue Apr  9 06:33:16 2013
@@ -1,8 +1,6 @@
-PREHOOK: query: -- small 2 part, 4 bucket & big 1 part, 2 bucket
-CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE
+PREHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE
 PREHOOK: type: CREATETABLE
-POSTHOOK: query: -- small 2 part, 4 bucket & big 1 part, 2 bucket
-CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE
+POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@bucket_small
 PREHOOK: query: load data local inpath '../data/files/srcsortbucket1outof4.txt' INTO TABLE bucket_small partition(ds='2008-04-08')

Modified: hive/trunk/ql/src/test/results/clientpositive/bucketcontext_5.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/bucketcontext_5.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/bucketcontext_5.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/bucketcontext_5.q.out Tue Apr  9 06:33:16 2013
@@ -1,8 +1,6 @@
-PREHOOK: query: -- small no part, 4 bucket & big no part, 2 bucket
-CREATE TABLE bucket_small (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE
+PREHOOK: query: CREATE TABLE bucket_small (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE
 PREHOOK: type: CREATETABLE
-POSTHOOK: query: -- small no part, 4 bucket & big no part, 2 bucket
-CREATE TABLE bucket_small (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE
+POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@bucket_small
 PREHOOK: query: load data local inpath '../data/files/srcsortbucket1outof4.txt' INTO TABLE bucket_small

Modified: hive/trunk/ql/src/test/results/clientpositive/bucketcontext_6.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/bucketcontext_6.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/bucketcontext_6.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/bucketcontext_6.q.out Tue Apr  9 06:33:16 2013
@@ -1,8 +1,6 @@
-PREHOOK: query: -- small no part, 4 bucket & big 2 part, 2 bucket
-CREATE TABLE bucket_small (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE
+PREHOOK: query: CREATE TABLE bucket_small (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE
 PREHOOK: type: CREATETABLE
-POSTHOOK: query: -- small no part, 4 bucket & big 2 part, 2 bucket
-CREATE TABLE bucket_small (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE
+POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@bucket_small
 PREHOOK: query: load data local inpath '../data/files/srcsortbucket1outof4.txt' INTO TABLE bucket_small

Modified: hive/trunk/ql/src/test/results/clientpositive/bucketcontext_7.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/bucketcontext_7.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/bucketcontext_7.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/bucketcontext_7.q.out Tue Apr  9 06:33:16 2013
@@ -1,8 +1,6 @@
-PREHOOK: query: -- small 2 part, 4 bucket & big 2 part, 2 bucket
-CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE
+PREHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE
 PREHOOK: type: CREATETABLE
-POSTHOOK: query: -- small 2 part, 4 bucket & big 2 part, 2 bucket
-CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE
+POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@bucket_small
 PREHOOK: query: load data local inpath '../data/files/srcsortbucket1outof4.txt' INTO TABLE bucket_small partition(ds='2008-04-08')

Modified: hive/trunk/ql/src/test/results/clientpositive/bucketcontext_8.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/bucketcontext_8.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/bucketcontext_8.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/bucketcontext_8.q.out Tue Apr  9 06:33:16 2013
@@ -1,8 +1,6 @@
-PREHOOK: query: -- small 2 part, 2 bucket & big 2 part, 4 bucket
-CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
+PREHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
 PREHOOK: type: CREATETABLE
-POSTHOOK: query: -- small 2 part, 2 bucket & big 2 part, 4 bucket
-CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
+POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@bucket_small
 PREHOOK: query: load data local inpath '../data/files/srcsortbucket1outof4.txt' INTO TABLE bucket_small partition(ds='2008-04-08')

Modified: hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin1.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin1.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin1.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin1.q.out Tue Apr  9 06:33:16 2013
@@ -13,14 +13,12 @@ PREHOOK: type: CREATETABLE
 POSTHOOK: query: CREATE TABLE srcbucket_mapjoin_part_2 (key int, value string) partitioned by (ds string) CLUSTERED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@srcbucket_mapjoin_part_2
-PREHOOK: query: -- empty partitions (HIVE-3205)
-explain extended
+PREHOOK: query: explain extended
 select /*+mapjoin(b)*/ a.key, a.value, b.value
 from srcbucket_mapjoin_part a join srcbucket_mapjoin_part_2 b
 on a.key=b.key where b.ds="2008-04-08"
 PREHOOK: type: QUERY
-POSTHOOK: query: -- empty partitions (HIVE-3205)
-explain extended
+POSTHOOK: query: explain extended
 select /*+mapjoin(b)*/ a.key, a.value, b.value
 from srcbucket_mapjoin_part a join srcbucket_mapjoin_part_2 b
 on a.key=b.key where b.ds="2008-04-08"

Modified: hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin10.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin10.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin10.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin10.q.out Tue Apr  9 06:33:16 2013
@@ -100,16 +100,12 @@ POSTHOOK: query: ALTER TABLE srcbucket_m
 POSTHOOK: type: ALTERTABLE_CLUSTER_SORT
 POSTHOOK: Input: default@srcbucket_mapjoin_part_2
 POSTHOOK: Output: default@srcbucket_mapjoin_part_2
-PREHOOK: query: -- The table bucketing metadata matches but the partition metadata does not, bucket map join should not be used
-
-EXPLAIN EXTENDED
+PREHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_2 b
 ON a.key = b.key AND a.part IS NOT NULL AND b.part IS NOT NULL
 PREHOOK: type: QUERY
-POSTHOOK: query: -- The table bucketing metadata matches but the partition metadata does not, bucket map join should not be used
-
-EXPLAIN EXTENDED
+POSTHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_2 b
 ON a.key = b.key AND a.part IS NOT NULL AND b.part IS NOT NULL

Modified: hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin11.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin11.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin11.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin11.q.out Tue Apr  9 06:33:16 2013
@@ -104,18 +104,12 @@ PREHOOK: Output: default@srcbucket_mapjo
 POSTHOOK: query: LOAD DATA LOCAL INPATH '../data/files/srcbucket21.txt' INTO TABLE srcbucket_mapjoin_part_2 PARTITION (part='2')
 POSTHOOK: type: LOAD
 POSTHOOK: Output: default@srcbucket_mapjoin_part_2@part=2
-PREHOOK: query: -- The table and partition bucketing metadata doesn't match but the bucket numbers of all partitions is
--- a power of 2 and the bucketing columns match so bucket map join should be used
-
-EXPLAIN EXTENDED
+PREHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_2 b
 ON a.key = b.key AND a.part IS NOT NULL AND b.part IS NOT NULL
 PREHOOK: type: QUERY
-POSTHOOK: query: -- The table and partition bucketing metadata doesn't match but the bucket numbers of all partitions is
--- a power of 2 and the bucketing columns match so bucket map join should be used
-
-EXPLAIN EXTENDED
+POSTHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_2 b
 ON a.key = b.key AND a.part IS NOT NULL AND b.part IS NOT NULL

Modified: hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin12.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin12.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin12.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin12.q.out Tue Apr  9 06:33:16 2013
@@ -74,16 +74,12 @@ POSTHOOK: query: ALTER TABLE srcbucket_m
 POSTHOOK: type: ALTERTABLE_CLUSTER_SORT
 POSTHOOK: Input: default@srcbucket_mapjoin_part_3
 POSTHOOK: Output: default@srcbucket_mapjoin_part_3
-PREHOOK: query: -- The partition bucketing metadata match but one table is not bucketed, bucket map join should still be used
-
-EXPLAIN EXTENDED
+PREHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_2 b
 ON a.key = b.key AND a.part = '1' and b.part = '1'
 PREHOOK: type: QUERY
-POSTHOOK: query: -- The partition bucketing metadata match but one table is not bucketed, bucket map join should still be used
-
-EXPLAIN EXTENDED
+POSTHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_2 b
 ON a.key = b.key AND a.part = '1' and b.part = '1'
@@ -266,16 +262,12 @@ POSTHOOK: Input: default@srcbucket_mapjo
 POSTHOOK: Input: default@srcbucket_mapjoin_part_2@part=1
 #### A masked pattern was here ####
 464
-PREHOOK: query: -- The table bucketing metadata match but one partition is not bucketed, bucket map join should not be used
-
-EXPLAIN EXTENDED
+PREHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_3 b
 ON a.key = b.key AND a.part = '1' and b.part = '1'
 PREHOOK: type: QUERY
-POSTHOOK: query: -- The table bucketing metadata match but one partition is not bucketed, bucket map join should not be used
-
-EXPLAIN EXTENDED
+POSTHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_3 b
 ON a.key = b.key AND a.part = '1' and b.part = '1'

Modified: hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin13.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin13.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin13.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin13.q.out Tue Apr  9 06:33:16 2013
@@ -5,14 +5,12 @@ POSTHOOK: query: CREATE TABLE srcbucket_
 CLUSTERED BY (value) INTO 2 BUCKETS
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@srcbucket_mapjoin_part_1
-PREHOOK: query: -- part=1 partition for srcbucket_mapjoin_part_1 is bucketed by 'value'
-INSERT OVERWRITE TABLE srcbucket_mapjoin_part_1 PARTITION (part='1')
+PREHOOK: query: INSERT OVERWRITE TABLE srcbucket_mapjoin_part_1 PARTITION (part='1')
 SELECT * FROM src
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
 PREHOOK: Output: default@srcbucket_mapjoin_part_1@part=1
-POSTHOOK: query: -- part=1 partition for srcbucket_mapjoin_part_1 is bucketed by 'value'
-INSERT OVERWRITE TABLE srcbucket_mapjoin_part_1 PARTITION (part='1')
+POSTHOOK: query: INSERT OVERWRITE TABLE srcbucket_mapjoin_part_1 PARTITION (part='1')
 SELECT * FROM src
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
@@ -29,14 +27,12 @@ POSTHOOK: Input: default@srcbucket_mapjo
 POSTHOOK: Output: default@srcbucket_mapjoin_part_1
 POSTHOOK: Lineage: srcbucket_mapjoin_part_1 PARTITION(part=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: srcbucket_mapjoin_part_1 PARTITION(part=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-PREHOOK: query: -- part=2 partition for srcbucket_mapjoin_part_1 is bucketed by 'key'
-INSERT OVERWRITE TABLE srcbucket_mapjoin_part_1 PARTITION (part='2')
+PREHOOK: query: INSERT OVERWRITE TABLE srcbucket_mapjoin_part_1 PARTITION (part='2')
 SELECT * FROM src
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
 PREHOOK: Output: default@srcbucket_mapjoin_part_1@part=2
-POSTHOOK: query: -- part=2 partition for srcbucket_mapjoin_part_1 is bucketed by 'key'
-INSERT OVERWRITE TABLE srcbucket_mapjoin_part_1 PARTITION (part='2')
+POSTHOOK: query: INSERT OVERWRITE TABLE srcbucket_mapjoin_part_1 PARTITION (part='2')
 SELECT * FROM src
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
@@ -56,14 +52,12 @@ POSTHOOK: Lineage: srcbucket_mapjoin_par
 POSTHOOK: Lineage: srcbucket_mapjoin_part_1 PARTITION(part=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
 POSTHOOK: Lineage: srcbucket_mapjoin_part_1 PARTITION(part=2).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: srcbucket_mapjoin_part_1 PARTITION(part=2).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-PREHOOK: query: -- part=1 partition for srcbucket_mapjoin_part_2 is bucketed by 'key'
-INSERT OVERWRITE TABLE srcbucket_mapjoin_part_2 PARTITION (part='1')
+PREHOOK: query: INSERT OVERWRITE TABLE srcbucket_mapjoin_part_2 PARTITION (part='1')
 SELECT * FROM src
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
 PREHOOK: Output: default@srcbucket_mapjoin_part_2@part=1
-POSTHOOK: query: -- part=1 partition for srcbucket_mapjoin_part_2 is bucketed by 'key'
-INSERT OVERWRITE TABLE srcbucket_mapjoin_part_2 PARTITION (part='1')
+POSTHOOK: query: INSERT OVERWRITE TABLE srcbucket_mapjoin_part_2 PARTITION (part='1')
 SELECT * FROM src
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
@@ -74,16 +68,12 @@ POSTHOOK: Lineage: srcbucket_mapjoin_par
 POSTHOOK: Lineage: srcbucket_mapjoin_part_1 PARTITION(part=2).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
 POSTHOOK: Lineage: srcbucket_mapjoin_part_2 PARTITION(part=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: srcbucket_mapjoin_part_2 PARTITION(part=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-PREHOOK: query: -- part=1 partition for srcbucket_mapjoin_part_1 is bucketed by 'value'
--- and it is also being joined. So, bucketed map-join cannot be performed
-EXPLAIN EXTENDED
+PREHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_2 b
 ON a.key = b.key
 PREHOOK: type: QUERY
-POSTHOOK: query: -- part=1 partition for srcbucket_mapjoin_part_1 is bucketed by 'value'
--- and it is also being joined. So, bucketed map-join cannot be performed
-EXPLAIN EXTENDED
+POSTHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_2 b
 ON a.key = b.key
@@ -320,16 +310,12 @@ POSTHOOK: Lineage: srcbucket_mapjoin_par
 POSTHOOK: Lineage: srcbucket_mapjoin_part_2 PARTITION(part=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: srcbucket_mapjoin_part_2 PARTITION(part=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
 2056
-PREHOOK: query: -- part=2 partition for srcbucket_mapjoin_part_1 is bucketed by 'key'
--- and it is being joined. So, bucketed map-join can be performed
-EXPLAIN EXTENDED
+PREHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_2 b
 ON a.key = b.key and a.part = '2'
 PREHOOK: type: QUERY
-POSTHOOK: query: -- part=2 partition for srcbucket_mapjoin_part_1 is bucketed by 'key'
--- and it is being joined. So, bucketed map-join can be performed
-EXPLAIN EXTENDED
+POSTHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_2 b
 ON a.key = b.key and a.part = '2'
@@ -538,16 +524,12 @@ POSTHOOK: Lineage: srcbucket_mapjoin_par
 POSTHOOK: Lineage: srcbucket_mapjoin_part_1 PARTITION(part=2).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
 POSTHOOK: Lineage: srcbucket_mapjoin_part_2 PARTITION(part=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: srcbucket_mapjoin_part_2 PARTITION(part=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-PREHOOK: query: -- part=2 partition for srcbucket_mapjoin_part_1 is bucketed by 'key'
--- and it is being joined. So, bucketed map-join can be performed
-EXPLAIN EXTENDED
+PREHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_2 b
 ON a.key = b.key
 PREHOOK: type: QUERY
-POSTHOOK: query: -- part=2 partition for srcbucket_mapjoin_part_1 is bucketed by 'key'
--- and it is being joined. So, bucketed map-join can be performed
-EXPLAIN EXTENDED
+POSTHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_2 b
 ON a.key = b.key
@@ -756,18 +738,12 @@ POSTHOOK: Lineage: srcbucket_mapjoin_par
 POSTHOOK: Lineage: srcbucket_mapjoin_part_1 PARTITION(part=2).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
 POSTHOOK: Lineage: srcbucket_mapjoin_part_2 PARTITION(part=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: srcbucket_mapjoin_part_2 PARTITION(part=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-PREHOOK: query: -- part=2 partition for srcbucket_mapjoin_part_1 is bucketed by 'key'
--- and it is being joined. So, bucketed map-join can be performed
--- The fact that the table is being bucketed by 'value' does not matter
-EXPLAIN EXTENDED
+PREHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_2 b
 ON a.key = b.key
 PREHOOK: type: QUERY
-POSTHOOK: query: -- part=2 partition for srcbucket_mapjoin_part_1 is bucketed by 'key'
--- and it is being joined. So, bucketed map-join can be performed
--- The fact that the table is being bucketed by 'value' does not matter
-EXPLAIN EXTENDED
+POSTHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_2 b
 ON a.key = b.key

Modified: hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin2.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin2.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin2.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin2.q.out Tue Apr  9 06:33:16 2013
@@ -1167,12 +1167,10 @@ POSTHOOK: Lineage: bucketmapjoin_tmp_res
 POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part_2)b.FieldSchema(name:value, type:string, comment:null), ]
 POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part_2)b.FieldSchema(name:value, type:string, comment:null), ]
 0	0	0
-PREHOOK: query: -- HIVE-3210
-load data local inpath '../data/files/srcbucket22.txt' INTO TABLE srcbucket_mapjoin_part_2 partition(ds='2008-04-09')
+PREHOOK: query: load data local inpath '../data/files/srcbucket22.txt' INTO TABLE srcbucket_mapjoin_part_2 partition(ds='2008-04-09')
 PREHOOK: type: LOAD
 PREHOOK: Output: default@srcbucket_mapjoin_part_2
-POSTHOOK: query: -- HIVE-3210
-load data local inpath '../data/files/srcbucket22.txt' INTO TABLE srcbucket_mapjoin_part_2 partition(ds='2008-04-09')
+POSTHOOK: query: load data local inpath '../data/files/srcbucket22.txt' INTO TABLE srcbucket_mapjoin_part_2 partition(ds='2008-04-09')
 POSTHOOK: type: LOAD
 POSTHOOK: Output: default@srcbucket_mapjoin_part_2
 POSTHOOK: Output: default@srcbucket_mapjoin_part_2@ds=2008-04-09

Modified: hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin8.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin8.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin8.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin8.q.out Tue Apr  9 06:33:16 2013
@@ -46,16 +46,12 @@ POSTHOOK: query: ALTER TABLE srcbucket_m
 POSTHOOK: type: ALTERTABLE_CLUSTER_SORT
 POSTHOOK: Input: default@srcbucket_mapjoin_part_2
 POSTHOOK: Output: default@srcbucket_mapjoin_part_2
-PREHOOK: query: -- The partition bucketing metadata match but the tables have different numbers of buckets, bucket map join should still be used
-
-EXPLAIN EXTENDED
+PREHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_2 b
 ON a.key = b.key AND a.part = '1' and b.part = '1'
 PREHOOK: type: QUERY
-POSTHOOK: query: -- The partition bucketing metadata match but the tables have different numbers of buckets, bucket map join should still be used
-
-EXPLAIN EXTENDED
+POSTHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_2 b
 ON a.key = b.key AND a.part = '1' and b.part = '1'
@@ -246,16 +242,12 @@ POSTHOOK: query: ALTER TABLE srcbucket_m
 POSTHOOK: type: ALTERTABLE_CLUSTER_SORT
 POSTHOOK: Input: default@srcbucket_mapjoin_part_2
 POSTHOOK: Output: default@srcbucket_mapjoin_part_2
-PREHOOK: query: -- The partition bucketing metadata match but the tables are bucketed on different columns, bucket map join should still be used
-
-EXPLAIN EXTENDED
+PREHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_2 b
 ON a.key = b.key AND a.part = '1' and b.part = '1'
 PREHOOK: type: QUERY
-POSTHOOK: query: -- The partition bucketing metadata match but the tables are bucketed on different columns, bucket map join should still be used
-
-EXPLAIN EXTENDED
+POSTHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_2 b
 ON a.key = b.key AND a.part = '1' and b.part = '1'

Modified: hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin9.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin9.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin9.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin9.q.out Tue Apr  9 06:33:16 2013
@@ -52,16 +52,12 @@ POSTHOOK: query: ALTER TABLE srcbucket_m
 POSTHOOK: type: ALTERTABLE_CLUSTER_SORT
 POSTHOOK: Input: default@srcbucket_mapjoin_part_2
 POSTHOOK: Output: default@srcbucket_mapjoin_part_2
-PREHOOK: query: -- The table bucketing metadata matches but the partitions have different numbers of buckets, bucket map join should not be used
-
-EXPLAIN EXTENDED
+PREHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_2 b
 ON a.key = b.key AND a.part = '1' and b.part = '1'
 PREHOOK: type: QUERY
-POSTHOOK: query: -- The table bucketing metadata matches but the partitions have different numbers of buckets, bucket map join should not be used
-
-EXPLAIN EXTENDED
+POSTHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_2 b
 ON a.key = b.key AND a.part = '1' and b.part = '1'
@@ -273,16 +269,12 @@ POSTHOOK: query: ALTER TABLE srcbucket_m
 POSTHOOK: type: ALTERTABLE_CLUSTER_SORT
 POSTHOOK: Input: default@srcbucket_mapjoin_part_2
 POSTHOOK: Output: default@srcbucket_mapjoin_part_2
-PREHOOK: query: -- The table bucketing metadata matches but the partitions are bucketed on different columns, bucket map join should not be used
-
-EXPLAIN EXTENDED
+PREHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_2 b
 ON a.key = b.key AND a.part = '1' AND b.part = '1'
 PREHOOK: type: QUERY
-POSTHOOK: query: -- The table bucketing metadata matches but the partitions are bucketed on different columns, bucket map join should not be used
-
-EXPLAIN EXTENDED
+POSTHOOK: query: EXPLAIN EXTENDED
 SELECT /*+ MAPJOIN(b) */ count(*)
 FROM srcbucket_mapjoin_part_1 a JOIN srcbucket_mapjoin_part_2 b
 ON a.key = b.key AND a.part = '1' AND b.part = '1'

Modified: hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out Tue Apr  9 06:33:16 2013
@@ -106,11 +106,9 @@ PREHOOK: Output: default@test4
 POSTHOOK: query: load data local inpath '../data/files/srcbucket22.txt' INTO TABLE test4
 POSTHOOK: type: LOAD
 POSTHOOK: Output: default@test4
-PREHOOK: query: -- should be allowed
-explain extended select /* + MAPJOIN(R) */ * from test1 L join test1 R on L.key=R.key AND L.value=R.value
+PREHOOK: query: explain extended select /* + MAPJOIN(R) */ * from test1 L join test1 R on L.key=R.key AND L.value=R.value
 PREHOOK: type: QUERY
-POSTHOOK: query: -- should be allowed
-explain extended select /* + MAPJOIN(R) */ * from test1 L join test1 R on L.key=R.key AND L.value=R.value
+POSTHOOK: query: explain extended select /* + MAPJOIN(R) */ * from test1 L join test1 R on L.key=R.key AND L.value=R.value
 POSTHOOK: type: QUERY
 ABSTRACT SYNTAX TREE:
   (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_TABREF (TOK_TABNAME test1) L) (TOK_TABREF (TOK_TABNAME test1) R) (AND (= (. (TOK_TABLE_OR_COL L) key) (. (TOK_TABLE_OR_COL R) key)) (= (. (TOK_TABLE_OR_COL L) value) (. (TOK_TABLE_OR_COL R) value))))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_HINTLIST (TOK_HINT TOK_MAPJOIN (TOK_HINTARGLIST R))) (TOK_SELEXPR TOK_ALLCOLREF))))
@@ -406,11 +404,9 @@ STAGE PLANS:
       limit: -1
 
 
-PREHOOK: query: -- should not apply bucket mapjoin
-explain extended select /* + MAPJOIN(R) */ * from test1 L join test1 R on L.key+L.key=R.key
+PREHOOK: query: explain extended select /* + MAPJOIN(R) */ * from test1 L join test1 R on L.key+L.key=R.key
 PREHOOK: type: QUERY
-POSTHOOK: query: -- should not apply bucket mapjoin
-explain extended select /* + MAPJOIN(R) */ * from test1 L join test1 R on L.key+L.key=R.key
+POSTHOOK: query: explain extended select /* + MAPJOIN(R) */ * from test1 L join test1 R on L.key+L.key=R.key
 POSTHOOK: type: QUERY
 ABSTRACT SYNTAX TREE:
   (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_TABREF (TOK_TABNAME test1) L) (TOK_TABREF (TOK_TABNAME test1) R) (= (+ (. (TOK_TABLE_OR_COL L) key) (. (TOK_TABLE_OR_COL L) key)) (. (TOK_TABLE_OR_COL R) key)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_HINTLIST (TOK_HINT TOK_MAPJOIN (TOK_HINTARGLIST R))) (TOK_SELEXPR TOK_ALLCOLREF))))

Modified: hive/trunk/ql/src/test/results/clientpositive/column_access_stats.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/column_access_stats.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/column_access_stats.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/column_access_stats.q.out Tue Apr  9 06:33:16 2013
@@ -1,6 +1,4 @@
-PREHOOK: query: -- This test is used for testing the ColumnAccessAnalyzer
-
-CREATE TABLE T1(key STRING, val STRING) STORED AS TEXTFILE
+PREHOOK: query: CREATE TABLE T1(key STRING, val STRING) STORED AS TEXTFILE
 PREHOOK: type: CREATETABLE
 PREHOOK: query: LOAD DATA LOCAL INPATH '../data/files/T1.txt' INTO TABLE T1
 PREHOOK: type: LOAD
@@ -9,8 +7,7 @@ PREHOOK: query: CREATE TABLE T2(key STRI
 PREHOOK: type: CREATETABLE
 PREHOOK: query: CREATE TABLE T3(key STRING, val STRING) STORED AS TEXTFILE
 PREHOOK: type: CREATETABLE
-PREHOOK: query: -- Simple select queries
-SELECT key FROM T1 ORDER BY key
+PREHOOK: query: SELECT key FROM T1 ORDER BY key
 PREHOOK: type: QUERY
 PREHOOK: Input: default@t1
 #### A masked pattern was here ####
@@ -46,8 +43,7 @@ PREHOOK: Input: default@t1
 1
 1
 1
-PREHOOK: query: -- More complicated select queries
-EXPLAIN SELECT key FROM (SELECT key, val FROM T1) subq1 ORDER BY key
+PREHOOK: query: EXPLAIN SELECT key FROM (SELECT key, val FROM T1) subq1 ORDER BY key
 PREHOOK: type: QUERY
 ABSTRACT SYNTAX TREE:
   (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME T1))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_TABLE_OR_COL val))))) subq1)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)))))
@@ -187,8 +183,7 @@ Columns:key,val
 24.0
 26.0
 36.0
-PREHOOK: query: -- Work with union
-EXPLAIN
+PREHOOK: query: EXPLAIN
 SELECT * FROM (
 SELECT key as c FROM T1
  UNION ALL
@@ -390,8 +385,7 @@ Columns:key
 8
 8
 8
-PREHOOK: query: -- Work with insert overwrite
-FROM T1
+PREHOOK: query: FROM T1
 INSERT OVERWRITE TABLE T2 SELECT key, count(1) GROUP BY key
 INSERT OVERWRITE TABLE T3 SELECT key, sum(val) GROUP BY key
 PREHOOK: type: QUERY
@@ -401,8 +395,7 @@ PREHOOK: Output: default@t3
 Table:default@t1
 Columns:key,val
 
-PREHOOK: query: -- Simple joins
-SELECT *
+PREHOOK: query: SELECT *
 FROM T1 JOIN T2
 ON T1.key = T2.key
 ORDER BY T1.key, T1.val, T2.key, T2.val
@@ -549,8 +542,7 @@ Columns:key,val
 Table:default@t1
 Columns:key,val
 
-PREHOOK: query: -- Map join
-SELECT /*+ MAPJOIN(a) */ * 
+PREHOOK: query: SELECT /*+ MAPJOIN(a) */ * 
 FROM T1 a JOIN T2 b 
 ON a.key = b.key
 ORDER BY a.key, a.val, b.key, b.val
@@ -570,8 +562,7 @@ Columns:key,val
 7	17	7	1
 8	18	8	2
 8	28	8	2
-PREHOOK: query: -- More joins
-EXPLAIN
+PREHOOK: query: EXPLAIN
 SELECT *
 FROM T1 JOIN T2
 ON T1.key = T2.key AND T1.val = 3 and T2.val = 3
@@ -846,8 +837,7 @@ Columns:key,val
 Table:default@t1
 Columns:key,val
 
-PREHOOK: query: -- Join followed by join
-EXPLAIN
+PREHOOK: query: EXPLAIN
 SELECT *
 FROM
 (

Modified: hive/trunk/ql/src/test/results/clientpositive/combine2_hadoop20.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/combine2_hadoop20.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/combine2_hadoop20.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/combine2_hadoop20.q.out Tue Apr  9 06:33:16 2013
@@ -2,28 +2,12 @@ PREHOOK: query: USE default
 PREHOOK: type: SWITCHDATABASE
 POSTHOOK: query: USE default
 POSTHOOK: type: SWITCHDATABASE
-PREHOOK: query: -- EXCLUDE_OS_WINDOWS
--- excluded on windows because of difference in file name encoding logic
-
-
-create table combine2(key string) partitioned by (value string)
+PREHOOK: query: create table combine2(key string) partitioned by (value string)
 PREHOOK: type: CREATETABLE
-POSTHOOK: query: -- EXCLUDE_OS_WINDOWS
--- excluded on windows because of difference in file name encoding logic
-
-
-create table combine2(key string) partitioned by (value string)
+POSTHOOK: query: create table combine2(key string) partitioned by (value string)
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@combine2
-PREHOOK: query: -- INCLUDE_HADOOP_MAJOR_VERSIONS(0.20, 0.20S)
--- This test sets mapred.max.split.size=256 and hive.merge.smallfiles.avgsize=0
--- in an attempt to force the generation of multiple splits and multiple output files.
--- However, Hadoop 0.20 is incapable of generating splits smaller than the block size
--- when using CombineFileInputFormat, so only one split is generated. This has a
--- significant impact on the results results of this test.
--- This issue was fixed in MAPREDUCE-2046 which is included in 0.22.
-
-insert overwrite table combine2 partition(value) 
+PREHOOK: query: insert overwrite table combine2 partition(value) 
 select * from (
    select key, value from src where key < 10
    union all 
@@ -33,15 +17,7 @@ select * from (
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
 PREHOOK: Output: default@combine2
-POSTHOOK: query: -- INCLUDE_HADOOP_MAJOR_VERSIONS(0.20, 0.20S)
--- This test sets mapred.max.split.size=256 and hive.merge.smallfiles.avgsize=0
--- in an attempt to force the generation of multiple splits and multiple output files.
--- However, Hadoop 0.20 is incapable of generating splits smaller than the block size
--- when using CombineFileInputFormat, so only one split is generated. This has a
--- significant impact on the results results of this test.
--- This issue was fixed in MAPREDUCE-2046 which is included in 0.22.
-
-insert overwrite table combine2 partition(value) 
+POSTHOOK: query: insert overwrite table combine2 partition(value) 
 select * from (
    select key, value from src where key < 10
    union all 

Modified: hive/trunk/ql/src/test/results/clientpositive/compute_stats_binary.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/compute_stats_binary.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/compute_stats_binary.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/compute_stats_binary.q.out Tue Apr  9 06:33:16 2013
@@ -3,12 +3,10 @@ PREHOOK: type: CREATETABLE
 POSTHOOK: query: create table tab_binary(a binary)
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@tab_binary
-PREHOOK: query: -- insert some data
-LOAD DATA LOCAL INPATH "../data/files/binary.txt" INTO TABLE tab_binary
+PREHOOK: query: LOAD DATA LOCAL INPATH "../data/files/binary.txt" INTO TABLE tab_binary
 PREHOOK: type: LOAD
 PREHOOK: Output: default@tab_binary
-POSTHOOK: query: -- insert some data
-LOAD DATA LOCAL INPATH "../data/files/binary.txt" INTO TABLE tab_binary
+POSTHOOK: query: LOAD DATA LOCAL INPATH "../data/files/binary.txt" INTO TABLE tab_binary
 POSTHOOK: type: LOAD
 POSTHOOK: Output: default@tab_binary
 PREHOOK: query: select count(*) from tab_binary
@@ -20,13 +18,11 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@tab_binary
 #### A masked pattern was here ####
 10
-PREHOOK: query: -- compute statistical summary of data
-select compute_stats(a, 16) from tab_binary
+PREHOOK: query: select compute_stats(a, 16) from tab_binary
 PREHOOK: type: QUERY
 PREHOOK: Input: default@tab_binary
 #### A masked pattern was here ####
-POSTHOOK: query: -- compute statistical summary of data
-select compute_stats(a, 16) from tab_binary
+POSTHOOK: query: select compute_stats(a, 16) from tab_binary
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@tab_binary
 #### A masked pattern was here ####

Modified: hive/trunk/ql/src/test/results/clientpositive/compute_stats_boolean.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/compute_stats_boolean.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/compute_stats_boolean.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/compute_stats_boolean.q.out Tue Apr  9 06:33:16 2013
@@ -3,12 +3,10 @@ PREHOOK: type: CREATETABLE
 POSTHOOK: query: create table tab_bool(a boolean)
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@tab_bool
-PREHOOK: query: -- insert some data
-LOAD DATA LOCAL INPATH "../data/files/bool.txt" INTO TABLE tab_bool
+PREHOOK: query: LOAD DATA LOCAL INPATH "../data/files/bool.txt" INTO TABLE tab_bool
 PREHOOK: type: LOAD
 PREHOOK: Output: default@tab_bool
-POSTHOOK: query: -- insert some data
-LOAD DATA LOCAL INPATH "../data/files/bool.txt" INTO TABLE tab_bool
+POSTHOOK: query: LOAD DATA LOCAL INPATH "../data/files/bool.txt" INTO TABLE tab_bool
 POSTHOOK: type: LOAD
 POSTHOOK: Output: default@tab_bool
 PREHOOK: query: select count(*) from tab_bool
@@ -20,13 +18,11 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@tab_bool
 #### A masked pattern was here ####
 33
-PREHOOK: query: -- compute statistical summary of data
-select compute_stats(a, 16) from tab_bool
+PREHOOK: query: select compute_stats(a, 16) from tab_bool
 PREHOOK: type: QUERY
 PREHOOK: Input: default@tab_bool
 #### A masked pattern was here ####
-POSTHOOK: query: -- compute statistical summary of data
-select compute_stats(a, 16) from tab_bool
+POSTHOOK: query: select compute_stats(a, 16) from tab_bool
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@tab_bool
 #### A masked pattern was here ####

Modified: hive/trunk/ql/src/test/results/clientpositive/compute_stats_double.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/compute_stats_double.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/compute_stats_double.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/compute_stats_double.q.out Tue Apr  9 06:33:16 2013
@@ -3,12 +3,10 @@ PREHOOK: type: CREATETABLE
 POSTHOOK: query: create table tab_double(a double)
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@tab_double
-PREHOOK: query: -- insert some data
-LOAD DATA LOCAL INPATH "../data/files/double.txt" INTO TABLE tab_double
+PREHOOK: query: LOAD DATA LOCAL INPATH "../data/files/double.txt" INTO TABLE tab_double
 PREHOOK: type: LOAD
 PREHOOK: Output: default@tab_double
-POSTHOOK: query: -- insert some data
-LOAD DATA LOCAL INPATH "../data/files/double.txt" INTO TABLE tab_double
+POSTHOOK: query: LOAD DATA LOCAL INPATH "../data/files/double.txt" INTO TABLE tab_double
 POSTHOOK: type: LOAD
 POSTHOOK: Output: default@tab_double
 PREHOOK: query: select count(*) from tab_double
@@ -20,13 +18,11 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@tab_double
 #### A masked pattern was here ####
 16
-PREHOOK: query: -- compute statistical summary of data
-select compute_stats(a, 16) from tab_double
+PREHOOK: query: select compute_stats(a, 16) from tab_double
 PREHOOK: type: QUERY
 PREHOOK: Input: default@tab_double
 #### A masked pattern was here ####
-POSTHOOK: query: -- compute statistical summary of data
-select compute_stats(a, 16) from tab_double
+POSTHOOK: query: select compute_stats(a, 16) from tab_double
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@tab_double
 #### A masked pattern was here ####

Modified: hive/trunk/ql/src/test/results/clientpositive/compute_stats_empty_table.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/compute_stats_empty_table.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/compute_stats_empty_table.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/compute_stats_empty_table.q.out Tue Apr  9 06:33:16 2013
@@ -12,13 +12,11 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@tab_empty
 #### A masked pattern was here ####
 0
-PREHOOK: query: -- compute statistical summary of data
-select compute_stats(a, 16) from tab_empty
+PREHOOK: query: select compute_stats(a, 16) from tab_empty
 PREHOOK: type: QUERY
 PREHOOK: Input: default@tab_empty
 #### A masked pattern was here ####
-POSTHOOK: query: -- compute statistical summary of data
-select compute_stats(a, 16) from tab_empty
+POSTHOOK: query: select compute_stats(a, 16) from tab_empty
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@tab_empty
 #### A masked pattern was here ####

Modified: hive/trunk/ql/src/test/results/clientpositive/compute_stats_long.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/compute_stats_long.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/compute_stats_long.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/compute_stats_long.q.out Tue Apr  9 06:33:16 2013
@@ -3,12 +3,10 @@ PREHOOK: type: CREATETABLE
 POSTHOOK: query: create table tab_int(a int)
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@tab_int
-PREHOOK: query: -- insert some data
-LOAD DATA LOCAL INPATH "../data/files/int.txt" INTO TABLE tab_int
+PREHOOK: query: LOAD DATA LOCAL INPATH "../data/files/int.txt" INTO TABLE tab_int
 PREHOOK: type: LOAD
 PREHOOK: Output: default@tab_int
-POSTHOOK: query: -- insert some data
-LOAD DATA LOCAL INPATH "../data/files/int.txt" INTO TABLE tab_int
+POSTHOOK: query: LOAD DATA LOCAL INPATH "../data/files/int.txt" INTO TABLE tab_int
 POSTHOOK: type: LOAD
 POSTHOOK: Output: default@tab_int
 PREHOOK: query: select count(*) from tab_int
@@ -20,13 +18,11 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@tab_int
 #### A masked pattern was here ####
 12
-PREHOOK: query: -- compute statistical summary of data
-select compute_stats(a, 16) from tab_int
+PREHOOK: query: select compute_stats(a, 16) from tab_int
 PREHOOK: type: QUERY
 PREHOOK: Input: default@tab_int
 #### A masked pattern was here ####
-POSTHOOK: query: -- compute statistical summary of data
-select compute_stats(a, 16) from tab_int
+POSTHOOK: query: select compute_stats(a, 16) from tab_int
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@tab_int
 #### A masked pattern was here ####

Modified: hive/trunk/ql/src/test/results/clientpositive/compute_stats_string.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/compute_stats_string.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/compute_stats_string.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/compute_stats_string.q.out Tue Apr  9 06:33:16 2013
@@ -3,12 +3,10 @@ PREHOOK: type: CREATETABLE
 POSTHOOK: query: create table tab_string(a string)
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@tab_string
-PREHOOK: query: -- insert some data
-LOAD DATA LOCAL INPATH "../data/files/string.txt" INTO TABLE tab_string
+PREHOOK: query: LOAD DATA LOCAL INPATH "../data/files/string.txt" INTO TABLE tab_string
 PREHOOK: type: LOAD
 PREHOOK: Output: default@tab_string
-POSTHOOK: query: -- insert some data
-LOAD DATA LOCAL INPATH "../data/files/string.txt" INTO TABLE tab_string
+POSTHOOK: query: LOAD DATA LOCAL INPATH "../data/files/string.txt" INTO TABLE tab_string
 POSTHOOK: type: LOAD
 POSTHOOK: Output: default@tab_string
 PREHOOK: query: select count(*) from tab_string
@@ -20,13 +18,11 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@tab_string
 #### A masked pattern was here ####
 10
-PREHOOK: query: -- compute statistical summary of data
-select compute_stats(a, 16) from tab_string
+PREHOOK: query: select compute_stats(a, 16) from tab_string
 PREHOOK: type: QUERY
 PREHOOK: Input: default@tab_string
 #### A masked pattern was here ####
-POSTHOOK: query: -- compute statistical summary of data
-select compute_stats(a, 16) from tab_string
+POSTHOOK: query: select compute_stats(a, 16) from tab_string
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@tab_string
 #### A masked pattern was here ####

Modified: hive/trunk/ql/src/test/results/clientpositive/convert_enum_to_string.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/convert_enum_to_string.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/convert_enum_to_string.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/convert_enum_to_string.q.out Tue Apr  9 06:33:16 2013
@@ -1,15 +1,11 @@
-PREHOOK: query: -- Ensure Enum fields are converted to strings (instead of struct<value:int>)
-
-create table convert_enum_to_string
+PREHOOK: query: create table convert_enum_to_string
   partitioned by (b string)
   row format serde "org.apache.hadoop.hive.serde2.thrift.ThriftDeserializer"
     with serdeproperties (
       "serialization.class"="org.apache.hadoop.hive.serde2.thrift.test.MegaStruct",
       "serialization.format"="org.apache.thrift.protocol.TBinaryProtocol")
 PREHOOK: type: CREATETABLE
-POSTHOOK: query: -- Ensure Enum fields are converted to strings (instead of struct<value:int>)
-
-create table convert_enum_to_string
+POSTHOOK: query: create table convert_enum_to_string
   partitioned by (b string)
   row format serde "org.apache.hadoop.hive.serde2.thrift.ThriftDeserializer"
     with serdeproperties (

Modified: hive/trunk/ql/src/test/results/clientpositive/create_alter_list_bucketing_table1.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/create_alter_list_bucketing_table1.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/create_alter_list_bucketing_table1.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/create_alter_list_bucketing_table1.q.out Tue Apr  9 06:33:16 2013
@@ -1,15 +1,7 @@
-PREHOOK: query: -- Test stored as directories
--- it covers a few cases
-
--- 1. create a table with stored as directories
-CREATE TABLE  if not exists stored_as_dirs_multiple (col1 STRING, col2 int, col3 STRING) 
+PREHOOK: query: CREATE TABLE  if not exists stored_as_dirs_multiple (col1 STRING, col2 int, col3 STRING) 
 SKEWED BY (col1, col2) ON (('s1',1), ('s3',3), ('s13',13), ('s78',78))  stored as DIRECTORIES
 PREHOOK: type: CREATETABLE
-POSTHOOK: query: -- Test stored as directories
--- it covers a few cases
-
--- 1. create a table with stored as directories
-CREATE TABLE  if not exists stored_as_dirs_multiple (col1 STRING, col2 int, col3 STRING) 
+POSTHOOK: query: CREATE TABLE  if not exists stored_as_dirs_multiple (col1 STRING, col2 int, col3 STRING) 
 SKEWED BY (col1, col2) ON (('s1',1), ('s3',3), ('s13',13), ('s78',78))  stored as DIRECTORIES
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@stored_as_dirs_multiple
@@ -46,13 +38,11 @@ Skewed Columns:     	[col1, col2]       
 Skewed Values:      	[[s1, 1], [s3, 3], [s13, 13], [s78, 78]]	 
 Storage Desc Params:	 	 
 	serialization.format	1                   
-PREHOOK: query: -- 2. turn off stored as directories but table is still a skewed table
-alter table stored_as_dirs_multiple not stored as DIRECTORIES
+PREHOOK: query: alter table stored_as_dirs_multiple not stored as DIRECTORIES
 PREHOOK: type: ALTERTABLE_SKEWED
 PREHOOK: Input: default@stored_as_dirs_multiple
 PREHOOK: Output: default@stored_as_dirs_multiple
-POSTHOOK: query: -- 2. turn off stored as directories but table is still a skewed table
-alter table stored_as_dirs_multiple not stored as DIRECTORIES
+POSTHOOK: query: alter table stored_as_dirs_multiple not stored as DIRECTORIES
 POSTHOOK: type: ALTERTABLE_SKEWED
 POSTHOOK: Input: default@stored_as_dirs_multiple
 POSTHOOK: Output: default@stored_as_dirs_multiple
@@ -88,13 +78,11 @@ Skewed Columns:     	[col1, col2]       
 Skewed Values:      	[[s1, 1], [s3, 3], [s13, 13], [s78, 78]]	 
 Storage Desc Params:	 	 
 	serialization.format	1                   
-PREHOOK: query: -- 3. turn off skewed
-alter table stored_as_dirs_multiple not skewed
+PREHOOK: query: alter table stored_as_dirs_multiple not skewed
 PREHOOK: type: ALTERTABLE_SKEWED
 PREHOOK: Input: default@stored_as_dirs_multiple
 PREHOOK: Output: default@stored_as_dirs_multiple
-POSTHOOK: query: -- 3. turn off skewed
-alter table stored_as_dirs_multiple not skewed
+POSTHOOK: query: alter table stored_as_dirs_multiple not skewed
 POSTHOOK: type: ALTERTABLE_SKEWED
 POSTHOOK: Input: default@stored_as_dirs_multiple
 POSTHOOK: Output: default@stored_as_dirs_multiple
@@ -128,11 +116,9 @@ Bucket Columns:     	[]                 
 Sort Columns:       	[]                  	 
 Storage Desc Params:	 	 
 	serialization.format	1                   
-PREHOOK: query: -- 4. alter a table to stored as directories
-CREATE TABLE stored_as_dirs_single (key STRING, value STRING)
+PREHOOK: query: CREATE TABLE stored_as_dirs_single (key STRING, value STRING)
 PREHOOK: type: CREATETABLE
-POSTHOOK: query: -- 4. alter a table to stored as directories
-CREATE TABLE stored_as_dirs_single (key STRING, value STRING)
+POSTHOOK: query: CREATE TABLE stored_as_dirs_single (key STRING, value STRING)
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@stored_as_dirs_single
 PREHOOK: query: alter table stored_as_dirs_single SKEWED BY (key) ON ('1','5','6') 
@@ -177,13 +163,11 @@ Skewed Columns:     	[key]              
 Skewed Values:      	[[1], [5], [6]]     	 
 Storage Desc Params:	 	 
 	serialization.format	1                   
-PREHOOK: query: -- 5. turn off skewed should turn off stored as directories too
-alter table stored_as_dirs_single not skewed
+PREHOOK: query: alter table stored_as_dirs_single not skewed
 PREHOOK: type: ALTERTABLE_SKEWED
 PREHOOK: Input: default@stored_as_dirs_single
 PREHOOK: Output: default@stored_as_dirs_single
-POSTHOOK: query: -- 5. turn off skewed should turn off stored as directories too
-alter table stored_as_dirs_single not skewed
+POSTHOOK: query: alter table stored_as_dirs_single not skewed
 POSTHOOK: type: ALTERTABLE_SKEWED
 POSTHOOK: Input: default@stored_as_dirs_single
 POSTHOOK: Output: default@stored_as_dirs_single
@@ -216,14 +200,12 @@ Bucket Columns:     	[]                 
 Sort Columns:       	[]                  	 
 Storage Desc Params:	 	 
 	serialization.format	1                   
-PREHOOK: query: -- 6. turn on stored as directories again
-alter table stored_as_dirs_single SKEWED BY (key) ON ('1','5','6') 
+PREHOOK: query: alter table stored_as_dirs_single SKEWED BY (key) ON ('1','5','6') 
 stored as DIRECTORIES
 PREHOOK: type: ALTERTABLE_SKEWED
 PREHOOK: Input: default@stored_as_dirs_single
 PREHOOK: Output: default@stored_as_dirs_single
-POSTHOOK: query: -- 6. turn on stored as directories again
-alter table stored_as_dirs_single SKEWED BY (key) ON ('1','5','6') 
+POSTHOOK: query: alter table stored_as_dirs_single SKEWED BY (key) ON ('1','5','6') 
 stored as DIRECTORIES
 POSTHOOK: type: ALTERTABLE_SKEWED
 POSTHOOK: Input: default@stored_as_dirs_single
@@ -260,11 +242,9 @@ Skewed Columns:     	[key]              
 Skewed Values:      	[[1], [5], [6]]     	 
 Storage Desc Params:	 	 
 	serialization.format	1                   
-PREHOOK: query: -- 7. create table like
-create table stored_as_dirs_single_like like stored_as_dirs_single
+PREHOOK: query: create table stored_as_dirs_single_like like stored_as_dirs_single
 PREHOOK: type: CREATETABLE
-POSTHOOK: query: -- 7. create table like
-create table stored_as_dirs_single_like like stored_as_dirs_single
+POSTHOOK: query: create table stored_as_dirs_single_like like stored_as_dirs_single
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@stored_as_dirs_single_like
 PREHOOK: query: describe formatted stored_as_dirs_single_like
@@ -299,13 +279,11 @@ Skewed Columns:     	[key]              
 Skewed Values:      	[[1], [5], [6]]     	 
 Storage Desc Params:	 	 
 	serialization.format	1                   
-PREHOOK: query: -- cleanup
-drop table stored_as_dirs_single
+PREHOOK: query: drop table stored_as_dirs_single
 PREHOOK: type: DROPTABLE
 PREHOOK: Input: default@stored_as_dirs_single
 PREHOOK: Output: default@stored_as_dirs_single
-POSTHOOK: query: -- cleanup
-drop table stored_as_dirs_single
+POSTHOOK: query: drop table stored_as_dirs_single
 POSTHOOK: type: DROPTABLE
 POSTHOOK: Input: default@stored_as_dirs_single
 POSTHOOK: Output: default@stored_as_dirs_single

Modified: hive/trunk/ql/src/test/results/clientpositive/create_big_view.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/create_big_view.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/create_big_view.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/create_big_view.q.out Tue Apr  9 06:33:16 2013
@@ -2,9 +2,7 @@ PREHOOK: query: DROP VIEW big_view
 PREHOOK: type: DROPVIEW
 POSTHOOK: query: DROP VIEW big_view
 POSTHOOK: type: DROPVIEW
-PREHOOK: query: -- Define a view with long SQL text to test metastore and other limits.
-
-CREATE VIEW big_view AS SELECT 
+PREHOOK: query: CREATE VIEW big_view AS SELECT 
 'aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa' AS a,
 'aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa',
 'aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa',
@@ -242,9 +240,7 @@ CREATE VIEW big_view AS SELECT 
 'aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa'
 FROM src
 PREHOOK: type: CREATEVIEW
-POSTHOOK: query: -- Define a view with long SQL text to test metastore and other limits.
-
-CREATE VIEW big_view AS SELECT 
+POSTHOOK: query: CREATE VIEW big_view AS SELECT 
 'aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa' AS a,
 'aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa',
 'aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa',

Modified: hive/trunk/ql/src/test/results/clientpositive/create_like2.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/create_like2.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/create_like2.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/create_like2.q.out Tue Apr  9 06:33:16 2013
@@ -1,12 +1,6 @@
-PREHOOK: query: -- Tests the copying over of Table Parameters according to a HiveConf setting
--- when doing a CREATE TABLE LIKE.
-
-CREATE TABLE table1(a INT, b STRING)
+PREHOOK: query: CREATE TABLE table1(a INT, b STRING)
 PREHOOK: type: CREATETABLE
-POSTHOOK: query: -- Tests the copying over of Table Parameters according to a HiveConf setting
--- when doing a CREATE TABLE LIKE.
-
-CREATE TABLE table1(a INT, b STRING)
+POSTHOOK: query: CREATE TABLE table1(a INT, b STRING)
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@table1
 PREHOOK: query: ALTER TABLE table1 SET TBLPROPERTIES ('a'='1', 'b'='2', 'c'='3', 'd' = '4')

Modified: hive/trunk/ql/src/test/results/clientpositive/create_like_tbl_props.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/create_like_tbl_props.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/create_like_tbl_props.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/create_like_tbl_props.q.out Tue Apr  9 06:33:16 2013
@@ -1,10 +1,6 @@
-PREHOOK: query: -- Test that CREATE TABLE LIKE commands can take explicit table properties
-
-CREATE TABLE test_table LIKE src TBLPROPERTIES('key'='value')
+PREHOOK: query: CREATE TABLE test_table LIKE src TBLPROPERTIES('key'='value')
 PREHOOK: type: CREATETABLE
-POSTHOOK: query: -- Test that CREATE TABLE LIKE commands can take explicit table properties
-
-CREATE TABLE test_table LIKE src TBLPROPERTIES('key'='value')
+POSTHOOK: query: CREATE TABLE test_table LIKE src TBLPROPERTIES('key'='value')
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@test_table
 PREHOOK: query: DESC FORMATTED test_table
@@ -76,13 +72,9 @@ Bucket Columns:     	[]                 
 Sort Columns:       	[]                  	 
 Storage Desc Params:	 	 
 	serialization.format	1                   
-PREHOOK: query: -- Test that CREATE TABLE LIKE commands can take default and explicit table properties
-
-CREATE TABLE test_table2 LIKE src TBLPROPERTIES('key2' = 'value2')
+PREHOOK: query: CREATE TABLE test_table2 LIKE src TBLPROPERTIES('key2' = 'value2')
 PREHOOK: type: CREATETABLE
-POSTHOOK: query: -- Test that CREATE TABLE LIKE commands can take default and explicit table properties
-
-CREATE TABLE test_table2 LIKE src TBLPROPERTIES('key2' = 'value2')
+POSTHOOK: query: CREATE TABLE test_table2 LIKE src TBLPROPERTIES('key2' = 'value2')
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@test_table2
 PREHOOK: query: DESC FORMATTED test_table2
@@ -116,13 +108,9 @@ Bucket Columns:     	[]                 
 Sort Columns:       	[]                  	 
 Storage Desc Params:	 	 
 	serialization.format	1                   
-PREHOOK: query: -- Test that properties inherited are overwritten by explicitly set ones
-
-CREATE TABLE test_table3 LIKE test_table2 TBLPROPERTIES('key2' = 'value3')
+PREHOOK: query: CREATE TABLE test_table3 LIKE test_table2 TBLPROPERTIES('key2' = 'value3')
 PREHOOK: type: CREATETABLE
-POSTHOOK: query: -- Test that properties inherited are overwritten by explicitly set ones
-
-CREATE TABLE test_table3 LIKE test_table2 TBLPROPERTIES('key2' = 'value3')
+POSTHOOK: query: CREATE TABLE test_table3 LIKE test_table2 TBLPROPERTIES('key2' = 'value3')
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@test_table3
 PREHOOK: query: DESC FORMATTED test_table3

Modified: hive/trunk/ql/src/test/results/clientpositive/create_like_view.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/create_like_view.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/create_like_view.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/create_like_view.q.out Tue Apr  9 06:33:16 2013
@@ -239,11 +239,9 @@ POSTHOOK: Lineage: table1.a SIMPLE [(src
 POSTHOOK: Lineage: table1.b SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
 POSTHOOK: Lineage: table2.a SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: table2.b SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-PREHOOK: query: -- check partitions
-create view view1 partitioned on (ds, hr) as select * from srcpart
+PREHOOK: query: create view view1 partitioned on (ds, hr) as select * from srcpart
 PREHOOK: type: CREATEVIEW
-POSTHOOK: query: -- check partitions
-create view view1 partitioned on (ds, hr) as select * from srcpart
+POSTHOOK: query: create view view1 partitioned on (ds, hr) as select * from srcpart
 POSTHOOK: type: CREATEVIEW
 POSTHOOK: Output: default@view1
 POSTHOOK: Lineage: table1.a SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]

Modified: hive/trunk/ql/src/test/results/clientpositive/create_or_replace_view.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/create_or_replace_view.q.out?rev=1465903&r1=1465902&r2=1465903&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/create_or_replace_view.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/create_or_replace_view.q.out Tue Apr  9 06:33:16 2013
@@ -39,11 +39,9 @@ Sort Columns:       	[]                 
 # View Information	 	 
 View Original Text: 	select * from srcpart	 
 View Expanded Text: 	select `srcpart`.`key`, `srcpart`.`value`, `srcpart`.`ds`, `srcpart`.`hr` from `default`.`srcpart`	 
-PREHOOK: query: -- modifying definition of unpartitioned view
-create or replace view v partitioned on (ds, hr) as select * from srcpart
+PREHOOK: query: create or replace view v partitioned on (ds, hr) as select * from srcpart
 PREHOOK: type: CREATEVIEW
-POSTHOOK: query: -- modifying definition of unpartitioned view
-create or replace view v partitioned on (ds, hr) as select * from srcpart
+POSTHOOK: query: create or replace view v partitioned on (ds, hr) as select * from srcpart
 POSTHOOK: type: CREATEVIEW
 POSTHOOK: Output: default@v
 PREHOOK: query: alter view v add partition (ds='2008-04-08',hr='11')
@@ -125,11 +123,9 @@ POSTHOOK: query: show partitions v
 POSTHOOK: type: SHOWPARTITIONS
 ds=2008-04-08/hr=11
 ds=2008-04-08/hr=12
-PREHOOK: query: -- altering partitioned view 1
-create or replace view v partitioned on (ds, hr) as select value, ds, hr from srcpart
+PREHOOK: query: create or replace view v partitioned on (ds, hr) as select value, ds, hr from srcpart
 PREHOOK: type: CREATEVIEW
-POSTHOOK: query: -- altering partitioned view 1
-create or replace view v partitioned on (ds, hr) as select value, ds, hr from srcpart
+POSTHOOK: query: create or replace view v partitioned on (ds, hr) as select value, ds, hr from srcpart
 POSTHOOK: type: CREATEVIEW
 POSTHOOK: Output: default@v
 PREHOOK: query: select * from v where value='val_409' and ds='2008-04-08' and hr='11'
@@ -188,11 +184,9 @@ POSTHOOK: query: show partitions v
 POSTHOOK: type: SHOWPARTITIONS
 ds=2008-04-08/hr=11
 ds=2008-04-08/hr=12
-PREHOOK: query: -- altering partitioned view 2
-create or replace view v partitioned on (ds, hr) as select key, value, ds, hr from srcpart
+PREHOOK: query: create or replace view v partitioned on (ds, hr) as select key, value, ds, hr from srcpart
 PREHOOK: type: CREATEVIEW
-POSTHOOK: query: -- altering partitioned view 2
-create or replace view v partitioned on (ds, hr) as select key, value, ds, hr from srcpart
+POSTHOOK: query: create or replace view v partitioned on (ds, hr) as select key, value, ds, hr from srcpart
 POSTHOOK: type: CREATEVIEW
 POSTHOOK: Output: default@v
 PREHOOK: query: select * from v where value='val_409' and ds='2008-04-08' and hr='11'
@@ -260,11 +254,9 @@ POSTHOOK: query: drop view v
 POSTHOOK: type: DROPVIEW
 POSTHOOK: Input: default@v
 POSTHOOK: Output: default@v
-PREHOOK: query: -- updating to fix view with invalid definition
-create table srcpart_temp like srcpart
+PREHOOK: query: create table srcpart_temp like srcpart
 PREHOOK: type: CREATETABLE
-POSTHOOK: query: -- updating to fix view with invalid definition
-create table srcpart_temp like srcpart
+POSTHOOK: query: create table srcpart_temp like srcpart
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: default@srcpart_temp
 PREHOOK: query: create view v partitioned on (ds, hr) as select * from srcpart_temp
@@ -280,11 +272,9 @@ POSTHOOK: query: drop table srcpart_temp
 POSTHOOK: type: DROPTABLE
 POSTHOOK: Input: default@srcpart_temp
 POSTHOOK: Output: default@srcpart_temp
-PREHOOK: query: -- v is now invalid
-create or replace view v partitioned on (ds, hr) as select * from srcpart
+PREHOOK: query: create or replace view v partitioned on (ds, hr) as select * from srcpart
 PREHOOK: type: CREATEVIEW
-POSTHOOK: query: -- v is now invalid
-create or replace view v partitioned on (ds, hr) as select * from srcpart
+POSTHOOK: query: create or replace view v partitioned on (ds, hr) as select * from srcpart
 POSTHOOK: type: CREATEVIEW
 POSTHOOK: Output: default@v
 PREHOOK: query: describe formatted v