You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by sp...@apache.org on 2016/05/27 15:38:03 UTC
[42/48] hive git commit: HIVE-13549: Remove jdk version specific out
files from Hive2 (Mohit Sabharwal, reviewed by Sergio Pena)
HIVE-13549: Remove jdk version specific out files from Hive2 (Mohit Sabharwal, reviewed by Sergio Pena)
Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/9349b8e5
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/9349b8e5
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/9349b8e5
Branch: refs/heads/java8
Commit: 9349b8e551b5c06312900e70c43f06b10f8a370d
Parents: 77f5ee7
Author: Mohit Sabharwal <mo...@cloudera.com>
Authored: Fri May 20 11:14:13 2016 -0500
Committer: Sergio Pena <se...@cloudera.com>
Committed: Fri May 27 10:36:28 2016 -0500
----------------------------------------------------------------------
.../columnstats_partlvl_invalid_values.q | 1 -
.../clientpositive/authorization_explain.q | 1 -
ql/src/test/queries/clientpositive/avro_date.q | 1 -
.../clientpositive/avro_deserialize_map_null.q | 1 -
.../clientpositive/avro_nullable_fields.q | 1 -
.../queries/clientpositive/avro_timestamp.q | 1 -
.../clientpositive/cbo_rp_outer_join_ppr.q | 1 -
ql/src/test/queries/clientpositive/char_udf1.q | 1 -
ql/src/test/queries/clientpositive/input4.q | 1 -
ql/src/test/queries/clientpositive/join0.q | 1 -
.../queries/clientpositive/list_bucket_dml_10.q | 1 -
.../queries/clientpositive/list_bucket_dml_11.q | 1 -
.../queries/clientpositive/list_bucket_dml_12.q | 1 -
.../queries/clientpositive/list_bucket_dml_13.q | 1 -
.../queries/clientpositive/list_bucket_dml_2.q | 1 -
.../queries/clientpositive/list_bucket_dml_4.q | 1 -
.../queries/clientpositive/list_bucket_dml_5.q | 1 -
.../queries/clientpositive/list_bucket_dml_6.q | 1 -
.../queries/clientpositive/list_bucket_dml_8.q | 1 -
.../queries/clientpositive/list_bucket_dml_9.q | 1 -
.../queries/clientpositive/outer_join_ppr.q | 1 -
.../queries/clientpositive/parquet_map_null.q | 1 -
ql/src/test/queries/clientpositive/plan_json.q | 1 -
.../queries/clientpositive/stats_list_bucket.q | 1 -
ql/src/test/queries/clientpositive/str_to_map.q | 1 -
.../clientpositive/subquery_multiinsert.q | 1 -
.../clientpositive/subquery_notin_having.q | 1 -
.../test/queries/clientpositive/varchar_udf1.q | 1 -
.../clientpositive/vector_cast_constant.q | 1 -
...mnstats_partlvl_invalid_values.q.java1.7.out | 73 --
...mnstats_partlvl_invalid_values.q.java1.8.out | 73 --
.../columnstats_partlvl_invalid_values.q.out | 69 ++
.../authorization_explain.q.java1.7.out | 44 -
.../authorization_explain.q.java1.8.out | 47 -
.../clientpositive/authorization_explain.q.out | 40 +
.../clientpositive/avro_date.q.java1.7.out | 130 --
.../clientpositive/avro_date.q.java1.8.out | 130 --
.../test/results/clientpositive/avro_date.q.out | 126 ++
.../avro_deserialize_map_null.q.java1.7.out | 57 -
.../avro_deserialize_map_null.q.java1.8.out | 57 -
.../avro_deserialize_map_null.q.out | 55 +
.../avro_nullable_fields.q.java1.7.out | 179 ---
.../avro_nullable_fields.q.java1.8.out | 179 ---
.../clientpositive/avro_nullable_fields.q.out | 177 +++
.../clientpositive/avro_timestamp.q.java1.7.out | 134 ---
.../clientpositive/avro_timestamp.q.java1.8.out | 134 ---
.../results/clientpositive/avro_timestamp.q.out | 132 +++
.../cbo_rp_outer_join_ppr.q.java1.7.out | 693 -----------
.../clientpositive/cbo_rp_outer_join_ppr.q.out | 691 +++++++++++
.../clientpositive/char_udf1.q.java1.7.out | 463 --------
.../clientpositive/char_udf1.q.java1.8.out | 457 -------
.../test/results/clientpositive/char_udf1.q.out | 459 +++++++
.../results/clientpositive/input4.q.java1.7.out | 559 ---------
.../results/clientpositive/input4.q.java1.8.out | 559 ---------
ql/src/test/results/clientpositive/input4.q.out | 555 +++++++++
.../results/clientpositive/join0.q.java1.7.out | 240 ----
.../results/clientpositive/join0.q.java1.8.out | 240 ----
ql/src/test/results/clientpositive/join0.q.out | 238 ++++
.../list_bucket_dml_10.q.java1.7.out | 361 ------
.../list_bucket_dml_10.q.java1.8.out | 389 ------
.../clientpositive/list_bucket_dml_10.q.out | 359 ++++++
.../list_bucket_dml_11.q.java1.7.out | 329 -----
.../list_bucket_dml_11.q.java1.8.out | 424 -------
.../clientpositive/list_bucket_dml_11.q.out | 327 +++++
.../list_bucket_dml_12.q.java1.7.out | 426 -------
.../list_bucket_dml_12.q.java1.8.out | 596 ----------
.../clientpositive/list_bucket_dml_12.q.out | 424 +++++++
.../list_bucket_dml_13.q.java1.7.out | 337 ------
.../list_bucket_dml_13.q.java1.8.out | 439 -------
.../clientpositive/list_bucket_dml_13.q.out | 335 ++++++
.../list_bucket_dml_2.q.java1.7.out | 591 ---------
.../list_bucket_dml_2.q.java1.8.out | 692 -----------
.../clientpositive/list_bucket_dml_2.q.out | 589 +++++++++
.../list_bucket_dml_4.q.java1.7.out | 813 -------------
.../list_bucket_dml_4.q.java1.8.out | 915 --------------
.../clientpositive/list_bucket_dml_4.q.out | 811 +++++++++++++
.../list_bucket_dml_5.q.java1.7.out | 506 --------
.../list_bucket_dml_5.q.java1.8.out | 617 ----------
.../clientpositive/list_bucket_dml_5.q.out | 504 ++++++++
.../list_bucket_dml_6.q.java1.7.out | 1007 ----------------
.../list_bucket_dml_6.q.java1.8.out | 1119 ------------------
.../clientpositive/list_bucket_dml_6.q.out | 1005 ++++++++++++++++
.../list_bucket_dml_8.q.java1.7.out | 641 ----------
.../list_bucket_dml_8.q.java1.8.out | 712 -----------
.../clientpositive/list_bucket_dml_8.q.out | 639 ++++++++++
.../list_bucket_dml_9.q.java1.7.out | 813 -------------
.../list_bucket_dml_9.q.java1.8.out | 915 --------------
.../clientpositive/list_bucket_dml_9.q.out | 811 +++++++++++++
.../clientpositive/llap/join0.q.java1.7.out | 242 ----
.../clientpositive/llap/join0.q.java1.8.out | 242 ----
.../results/clientpositive/llap/join0.q.out | 243 ++++
.../llap/vector_cast_constant.q.java1.7.out | 217 ----
.../llap/vector_cast_constant.q.java1.8.out | 217 ----
.../llap/vector_cast_constant.q.out | 216 ++++
.../clientpositive/outer_join_ppr.q.java1.7.out | 685 -----------
.../clientpositive/outer_join_ppr.q.java1.8.out | 855 -------------
.../results/clientpositive/outer_join_ppr.q.out | 683 +++++++++++
.../parquet_map_null.q.java1.7.out | 70 --
.../parquet_map_null.q.java1.8.out | 70 --
.../clientpositive/parquet_map_null.q.out | 68 ++
.../clientpositive/plan_json.q.java1.7.out | 13 -
.../clientpositive/plan_json.q.java1.8.out | 13 -
.../test/results/clientpositive/plan_json.q.out | 11 +
.../clientpositive/spark/join0.q.java1.7.out | 238 ----
.../clientpositive/spark/join0.q.java1.8.out | 238 ----
.../results/clientpositive/spark/join0.q.out | 20 +-
.../spark/list_bucket_dml_10.q.java1.7.out | 252 ----
.../spark/list_bucket_dml_10.q.java1.8.out | 280 -----
.../spark/list_bucket_dml_10.q.out | 250 ++++
.../spark/list_bucket_dml_2.q.java1.7.out | 591 ---------
.../spark/list_bucket_dml_2.q.java1.8.out | 663 -----------
.../spark/list_bucket_dml_2.q.out | Bin 28667 -> 27128 bytes
.../spark/outer_join_ppr.q.java1.7.out | 709 -----------
.../spark/outer_join_ppr.q.java1.8.out | 879 --------------
.../clientpositive/spark/outer_join_ppr.q.out | 490 ++------
.../spark/subquery_multiinsert.q.java1.7.out | 886 --------------
.../spark/subquery_multiinsert.q.java1.8.out | 890 --------------
.../spark/subquery_multiinsert.q.out | 56 +-
.../spark/vector_cast_constant.q.java1.7.out | 217 ----
.../spark/vector_cast_constant.q.java1.8.out | 203 ----
.../spark/vector_cast_constant.q.out | 54 +-
.../stats_list_bucket.q.java1.7.out | 191 ---
.../stats_list_bucket.q.java1.8.out | 193 ---
.../clientpositive/stats_list_bucket.q.out | 189 +++
.../clientpositive/str_to_map.q.java1.7.out | 220 ----
.../clientpositive/str_to_map.q.java1.8.out | 219 ----
.../results/clientpositive/str_to_map.q.out | 216 ++++
.../subquery_multiinsert.q.java1.7.out | 999 ----------------
.../subquery_multiinsert.q.java1.8.out | 999 ----------------
.../clientpositive/subquery_multiinsert.q.out | 997 ++++++++++++++++
.../subquery_notin_having.q.java1.7.out | 766 ------------
.../subquery_notin_having.q.java1.8.out | 762 ------------
.../clientpositive/subquery_notin_having.q.out | 764 ++++++++++++
.../clientpositive/tez/join0.q.java1.7.out | 239 ----
.../clientpositive/tez/join0.q.java1.8.out | 236 ----
.../test/results/clientpositive/tez/join0.q.out | 237 ++++
.../tez/vector_cast_constant.q.java1.7.out | 218 ----
.../tez/vector_cast_constant.q.java1.8.out | 216 ----
.../tez/vector_cast_constant.q.out | 214 ++++
.../clientpositive/varchar_udf1.q.java1.7.out | 457 -------
.../clientpositive/varchar_udf1.q.java1.8.out | 457 -------
.../results/clientpositive/varchar_udf1.q.out | 453 +++++++
.../vector_cast_constant.q.java1.7.out | 220 ----
.../vector_cast_constant.q.java1.8.out | 197 ---
.../clientpositive/vector_cast_constant.q.out | 53 +-
145 files changed, 13120 insertions(+), 32818 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientnegative/columnstats_partlvl_invalid_values.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientnegative/columnstats_partlvl_invalid_values.q b/ql/src/test/queries/clientnegative/columnstats_partlvl_invalid_values.q
index 712ece7..8521631 100644
--- a/ql/src/test/queries/clientnegative/columnstats_partlvl_invalid_values.q
+++ b/ql/src/test/queries/clientnegative/columnstats_partlvl_invalid_values.q
@@ -1,4 +1,3 @@
--- JAVA_VERSION_SPECIFIC_OUTPUT
DROP TABLE Employee_Part;
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/authorization_explain.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/authorization_explain.q b/ql/src/test/queries/clientpositive/authorization_explain.q
index 6a9475c..d429704 100644
--- a/ql/src/test/queries/clientpositive/authorization_explain.q
+++ b/ql/src/test/queries/clientpositive/authorization_explain.q
@@ -2,7 +2,6 @@ set hive.security.authorization.manager=org.apache.hadoop.hive.ql.security.autho
set hive.mapred.mode=nonstrict;
set hive.security.authorization.enabled=true;
--- JAVA_VERSION_SPECIFIC_OUTPUT
explain authorization select * from src join srcpart;
explain formatted authorization select * from src join srcpart;
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/avro_date.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/avro_date.q b/ql/src/test/queries/clientpositive/avro_date.q
index 15c07de..7169822 100644
--- a/ql/src/test/queries/clientpositive/avro_date.q
+++ b/ql/src/test/queries/clientpositive/avro_date.q
@@ -1,5 +1,4 @@
set hive.mapred.mode=nonstrict;
--- JAVA_VERSION_SPECIFIC_OUTPUT
DROP TABLE avro_date_staging;
DROP TABLE avro_date;
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/avro_deserialize_map_null.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/avro_deserialize_map_null.q b/ql/src/test/queries/clientpositive/avro_deserialize_map_null.q
index 962e649..42258d9 100644
--- a/ql/src/test/queries/clientpositive/avro_deserialize_map_null.q
+++ b/ql/src/test/queries/clientpositive/avro_deserialize_map_null.q
@@ -4,7 +4,6 @@
-- fileSchema = [{ "type" : "map", "values" : ["string","null"]}, "null"]
-- recordSchema = ["null", { "type" : "map", "values" : ["string","null"]}]
--- JAVA_VERSION_SPECIFIC_OUTPUT
DROP TABLE IF EXISTS avro_table;
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/avro_nullable_fields.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/avro_nullable_fields.q b/ql/src/test/queries/clientpositive/avro_nullable_fields.q
index 9ba7441..cb398d6 100644
--- a/ql/src/test/queries/clientpositive/avro_nullable_fields.q
+++ b/ql/src/test/queries/clientpositive/avro_nullable_fields.q
@@ -1,6 +1,5 @@
-- Verify that nullable fields properly work
--- JAVA_VERSION_SPECIFIC_OUTPUT
CREATE TABLE test_serializer(string1 STRING,
int1 INT,
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/avro_timestamp.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/avro_timestamp.q b/ql/src/test/queries/clientpositive/avro_timestamp.q
index 7bf0dc8..847f250 100644
--- a/ql/src/test/queries/clientpositive/avro_timestamp.q
+++ b/ql/src/test/queries/clientpositive/avro_timestamp.q
@@ -1,7 +1,6 @@
set hive.mapred.mode=nonstrict;
-- Exclude test on Windows due to space character being escaped in Hive paths on Windows.
-- EXCLUDE_OS_WINDOWS
--- JAVA_VERSION_SPECIFIC_OUTPUT
DROP TABLE avro_timestamp_staging;
DROP TABLE avro_timestamp;
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/cbo_rp_outer_join_ppr.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/cbo_rp_outer_join_ppr.q b/ql/src/test/queries/clientpositive/cbo_rp_outer_join_ppr.q
index c497ce9..d8f726e 100644
--- a/ql/src/test/queries/clientpositive/cbo_rp_outer_join_ppr.q
+++ b/ql/src/test/queries/clientpositive/cbo_rp_outer_join_ppr.q
@@ -4,7 +4,6 @@ set hive.cbo.returnpath.hiveop=true;
set hive.optimize.ppd=true;
-- SORT_QUERY_RESULTS
--- JAVA_VERSION_SPECIFIC_OUTPUT
EXPLAIN EXTENDED
FROM
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/char_udf1.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/char_udf1.q b/ql/src/test/queries/clientpositive/char_udf1.q
index 09012b4..39aa0e0 100644
--- a/ql/src/test/queries/clientpositive/char_udf1.q
+++ b/ql/src/test/queries/clientpositive/char_udf1.q
@@ -4,7 +4,6 @@ create table char_udf_1 (c1 string, c2 string, c3 char(10), c4 char(20));
insert overwrite table char_udf_1
select key, value, key, value from src where key = '238' limit 1;
--- JAVA_VERSION_SPECIFIC_OUTPUT
-- UDFs with char support
select
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/input4.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/input4.q b/ql/src/test/queries/clientpositive/input4.q
index 83edbe2..90fcbdd 100644
--- a/ql/src/test/queries/clientpositive/input4.q
+++ b/ql/src/test/queries/clientpositive/input4.q
@@ -1,4 +1,3 @@
--- JAVA_VERSION_SPECIFIC_OUTPUT
CREATE TABLE INPUT4(KEY STRING, VALUE STRING) STORED AS TEXTFILE;
EXPLAIN
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/join0.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/join0.q b/ql/src/test/queries/clientpositive/join0.q
index 66f2ef3..3252847 100644
--- a/ql/src/test/queries/clientpositive/join0.q
+++ b/ql/src/test/queries/clientpositive/join0.q
@@ -1,6 +1,5 @@
set hive.mapred.mode=nonstrict;
set hive.explain.user=false;
--- JAVA_VERSION_SPECIFIC_OUTPUT
-- SORT_QUERY_RESULTS
EXPLAIN
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/list_bucket_dml_10.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_10.q b/ql/src/test/queries/clientpositive/list_bucket_dml_10.q
index 5d3dade..f25c174 100644
--- a/ql/src/test/queries/clientpositive/list_bucket_dml_10.q
+++ b/ql/src/test/queries/clientpositive/list_bucket_dml_10.q
@@ -1,7 +1,6 @@
set mapred.input.dir.recursive=true;
-- run this test case in minimr to ensure it works in cluster
--- JAVA_VERSION_SPECIFIC_OUTPUT
-- list bucketing DML: static partition. multiple skewed columns.
-- ds=2008-04-08/hr=11/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME:
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/list_bucket_dml_11.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_11.q b/ql/src/test/queries/clientpositive/list_bucket_dml_11.q
index 2d22d66..8ac1627 100644
--- a/ql/src/test/queries/clientpositive/list_bucket_dml_11.q
+++ b/ql/src/test/queries/clientpositive/list_bucket_dml_11.q
@@ -6,7 +6,6 @@ set hive.merge.mapredfiles=false;
-- Ensure it works if skewed column is not the first column in the table columns
-- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23)
--- JAVA_VERSION_SPECIFIC_OUTPUT
-- list bucketing DML: static partition. multiple skewed columns.
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/list_bucket_dml_12.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_12.q b/ql/src/test/queries/clientpositive/list_bucket_dml_12.q
index ac063cc..9facfa5 100644
--- a/ql/src/test/queries/clientpositive/list_bucket_dml_12.q
+++ b/ql/src/test/queries/clientpositive/list_bucket_dml_12.q
@@ -7,7 +7,6 @@ set hive.merge.mapredfiles=false;
-- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23)
-- SORT_QUERY_RESULTS
--- JAVA_VERSION_SPECIFIC_OUTPUT
-- test where the skewed values are more than 1 say columns no. 2 and 4 in a table with 5 columns
create table list_bucketing_mul_col (col1 String, col2 String, col3 String, col4 String, col5 string)
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/list_bucket_dml_13.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_13.q b/ql/src/test/queries/clientpositive/list_bucket_dml_13.q
index d68ca93..0fe7f61 100644
--- a/ql/src/test/queries/clientpositive/list_bucket_dml_13.q
+++ b/ql/src/test/queries/clientpositive/list_bucket_dml_13.q
@@ -7,7 +7,6 @@ set hive.merge.mapredfiles=false;
-- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23)
-- SORT_QUERY_RESULTS
--- JAVA_VERSION_SPECIFIC_OUTPUT
-- test where the skewed values are more than 1 say columns no. 2 and 4 in a table with 5 columns
create table list_bucketing_mul_col (col1 String, col2 String, col3 String, col4 String, col5 string)
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/list_bucket_dml_2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_2.q b/ql/src/test/queries/clientpositive/list_bucket_dml_2.q
index 263a002..c6dceab 100644
--- a/ql/src/test/queries/clientpositive/list_bucket_dml_2.q
+++ b/ql/src/test/queries/clientpositive/list_bucket_dml_2.q
@@ -10,7 +10,6 @@ set hive.stats.reliable=true;
-- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23)
-- SORT_QUERY_RESULTS
--- JAVA_VERSION_SPECIFIC_OUTPUT
-- list bucketing DML: static partition. multiple skewed columns.
-- ds=2008-04-08/hr=11/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME:
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/list_bucket_dml_4.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_4.q b/ql/src/test/queries/clientpositive/list_bucket_dml_4.q
index 86ff342..950409d 100644
--- a/ql/src/test/queries/clientpositive/list_bucket_dml_4.q
+++ b/ql/src/test/queries/clientpositive/list_bucket_dml_4.q
@@ -9,7 +9,6 @@ set hive.merge.mapredfiles=false;
-- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23)
-- SORT_QUERY_RESULTS
--- JAVA_VERSION_SPECIFIC_OUTPUT
-- list bucketing DML: static partition. multiple skewed columns. merge.
-- ds=2008-04-08/hr=11/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME:
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/list_bucket_dml_5.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_5.q b/ql/src/test/queries/clientpositive/list_bucket_dml_5.q
index ace7ba9..fce8e2e 100644
--- a/ql/src/test/queries/clientpositive/list_bucket_dml_5.q
+++ b/ql/src/test/queries/clientpositive/list_bucket_dml_5.q
@@ -10,7 +10,6 @@ set mapred.input.dir.recursive=true;
-- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23)
-- SORT_QUERY_RESULTS
--- JAVA_VERSION_SPECIFIC_OUTPUT
-- create a skewed table
create table list_bucketing_dynamic_part (key String, value String)
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/list_bucket_dml_6.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_6.q b/ql/src/test/queries/clientpositive/list_bucket_dml_6.q
index 5684788..631c938 100644
--- a/ql/src/test/queries/clientpositive/list_bucket_dml_6.q
+++ b/ql/src/test/queries/clientpositive/list_bucket_dml_6.q
@@ -47,7 +47,6 @@ set hive.merge.mapredfiles=false;
-- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23)
-- SORT_QUERY_RESULTS
--- JAVA_VERSION_SPECIFIC_OUTPUT
-- create a skewed table
create table list_bucketing_dynamic_part (key String, value String)
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/list_bucket_dml_8.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_8.q b/ql/src/test/queries/clientpositive/list_bucket_dml_8.q
index d904543..6d73896 100644
--- a/ql/src/test/queries/clientpositive/list_bucket_dml_8.q
+++ b/ql/src/test/queries/clientpositive/list_bucket_dml_8.q
@@ -48,7 +48,6 @@ set hive.merge.mapredfiles=false;
-- 118 000002_0
-- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23)
--- JAVA_VERSION_SPECIFIC_OUTPUT
-- create a skewed table
create table list_bucketing_dynamic_part (key String, value String)
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/list_bucket_dml_9.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_9.q b/ql/src/test/queries/clientpositive/list_bucket_dml_9.q
index 620750c..d2e24af 100644
--- a/ql/src/test/queries/clientpositive/list_bucket_dml_9.q
+++ b/ql/src/test/queries/clientpositive/list_bucket_dml_9.q
@@ -9,7 +9,6 @@ set hive.merge.mapredfiles=false;
-- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23)
-- SORT_QUERY_RESULTS
--- JAVA_VERSION_SPECIFIC_OUTPUT
-- list bucketing DML: static partition. multiple skewed columns. merge.
-- ds=2008-04-08/hr=11/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME:
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/outer_join_ppr.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/outer_join_ppr.q b/ql/src/test/queries/clientpositive/outer_join_ppr.q
index 497a4d1..60a06ae 100644
--- a/ql/src/test/queries/clientpositive/outer_join_ppr.q
+++ b/ql/src/test/queries/clientpositive/outer_join_ppr.q
@@ -2,7 +2,6 @@ set hive.mapred.mode=nonstrict;
set hive.optimize.ppd=true;
-- SORT_QUERY_RESULTS
--- JAVA_VERSION_SPECIFIC_OUTPUT
EXPLAIN EXTENDED
FROM
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/parquet_map_null.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/parquet_map_null.q b/ql/src/test/queries/clientpositive/parquet_map_null.q
index 61058f3..e154159 100644
--- a/ql/src/test/queries/clientpositive/parquet_map_null.q
+++ b/ql/src/test/queries/clientpositive/parquet_map_null.q
@@ -1,5 +1,4 @@
-- This test attempts to write a parquet table from an avro table that contains map null values
--- JAVA_VERSION_SPECIFIC_OUTPUT
DROP TABLE IF EXISTS avro_table;
DROP TABLE IF EXISTS parquet_table;
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/plan_json.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/plan_json.q b/ql/src/test/queries/clientpositive/plan_json.q
index aa2b134..503b55d 100644
--- a/ql/src/test/queries/clientpositive/plan_json.q
+++ b/ql/src/test/queries/clientpositive/plan_json.q
@@ -1,5 +1,4 @@
-- explain plan json: the query gets the formatted json output of the query plan of the hive query
--- JAVA_VERSION_SPECIFIC_OUTPUT
EXPLAIN FORMATTED SELECT count(1) FROM src;
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/stats_list_bucket.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/stats_list_bucket.q b/ql/src/test/queries/clientpositive/stats_list_bucket.q
index 51137a8..536702c 100644
--- a/ql/src/test/queries/clientpositive/stats_list_bucket.q
+++ b/ql/src/test/queries/clientpositive/stats_list_bucket.q
@@ -1,6 +1,5 @@
-- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23)
--- JAVA_VERSION_SPECIFIC_OUTPUT
drop table stats_list_bucket;
drop table stats_list_bucket_1;
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/str_to_map.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/str_to_map.q b/ql/src/test/queries/clientpositive/str_to_map.q
index f2993b1..3280d89 100644
--- a/ql/src/test/queries/clientpositive/str_to_map.q
+++ b/ql/src/test/queries/clientpositive/str_to_map.q
@@ -1,7 +1,6 @@
set hive.mapred.mode=nonstrict;
set hive.fetch.task.conversion=more;
--- JAVA_VERSION_SPECIFIC_OUTPUT
desc function str_to_map;
desc function extended str_to_map;
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/subquery_multiinsert.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/subquery_multiinsert.q b/ql/src/test/queries/clientpositive/subquery_multiinsert.q
index bea2e13..9d70f51 100644
--- a/ql/src/test/queries/clientpositive/subquery_multiinsert.q
+++ b/ql/src/test/queries/clientpositive/subquery_multiinsert.q
@@ -2,7 +2,6 @@ set hive.mapred.mode=nonstrict;
set hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.PostExecutePrinter,org.apache.hadoop.hive.ql.hooks.PrintCompletedTasksHook;
-- SORT_QUERY_RESULTS
--- JAVA_VERSION_SPECIFIC_OUTPUT
CREATE TABLE src_4(
key STRING,
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/subquery_notin_having.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/subquery_notin_having.q b/ql/src/test/queries/clientpositive/subquery_notin_having.q
index 8b2914d..05148df 100644
--- a/ql/src/test/queries/clientpositive/subquery_notin_having.q
+++ b/ql/src/test/queries/clientpositive/subquery_notin_having.q
@@ -1,6 +1,5 @@
set hive.mapred.mode=nonstrict;
-- non agg, non corr
--- JAVA_VERSION_SPECIFIC_OUTPUT
explain
select key, count(*)
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/varchar_udf1.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/varchar_udf1.q b/ql/src/test/queries/clientpositive/varchar_udf1.q
index ff40b31..4d1f884 100644
--- a/ql/src/test/queries/clientpositive/varchar_udf1.q
+++ b/ql/src/test/queries/clientpositive/varchar_udf1.q
@@ -4,7 +4,6 @@ create table varchar_udf_1 (c1 string, c2 string, c3 varchar(10), c4 varchar(20)
insert overwrite table varchar_udf_1
select key, value, key, value from src where key = '238' limit 1;
--- JAVA_VERSION_SPECIFIC_OUTPUT
-- UDFs with varchar support
select
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/queries/clientpositive/vector_cast_constant.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/vector_cast_constant.q b/ql/src/test/queries/clientpositive/vector_cast_constant.q
index c50dd8f..94bee09 100644
--- a/ql/src/test/queries/clientpositive/vector_cast_constant.q
+++ b/ql/src/test/queries/clientpositive/vector_cast_constant.q
@@ -2,7 +2,6 @@ set hive.mapred.mode=nonstrict;
set hive.explain.user=false;
SET hive.vectorized.execution.enabled=true;
--- JAVA_VERSION_SPECIFIC_OUTPUT
DROP TABLE over1k;
DROP TABLE over1korc;
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.java1.7.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.java1.7.out b/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.java1.7.out
deleted file mode 100644
index 4ea70e3..0000000
--- a/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.java1.7.out
+++ /dev/null
@@ -1,73 +0,0 @@
-PREHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT
-
-DROP TABLE Employee_Part
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT
-
-DROP TABLE Employee_Part
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE Employee_Part(employeeID int, employeeName String) partitioned by (employeeSalary double, country string)
-row format delimited fields terminated by '|' stored as textfile
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@Employee_Part
-POSTHOOK: query: CREATE TABLE Employee_Part(employeeID int, employeeName String) partitioned by (employeeSalary double, country string)
-row format delimited fields terminated by '|' stored as textfile
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@Employee_Part
-PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='USA')
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@employee_part
-POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='USA')
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@employee_part
-POSTHOOK: Output: default@employee_part@employeesalary=2000.0/country=USA
-PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='UK')
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@employee_part
-POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='UK')
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@employee_part
-POSTHOOK: Output: default@employee_part@employeesalary=2000.0/country=UK
-PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='USA')
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@employee_part
-POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='USA')
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@employee_part
-POSTHOOK: Output: default@employee_part@employeesalary=3000.0/country=USA
-PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='4000.0', country='USA')
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@employee_part
-POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='4000.0', country='USA')
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@employee_part
-POSTHOOK: Output: default@employee_part@employeesalary=4000.0/country=USA
-PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3500.0', country='UK')
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@employee_part
-POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3500.0', country='UK')
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@employee_part
-POSTHOOK: Output: default@employee_part@employeesalary=3500.0/country=UK
-PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='UK')
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@employee_part
-POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='UK')
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@employee_part
-POSTHOOK: Output: default@employee_part@employeesalary=3000.0/country=UK
-FAILED: SemanticException [Error 30007]: Invalid partitioning key/value specified in ANALYZE statement : {employeesalary=4000.0, country=Canada}
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.java1.8.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.java1.8.out b/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.java1.8.out
deleted file mode 100644
index 7cae55e..0000000
--- a/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.java1.8.out
+++ /dev/null
@@ -1,73 +0,0 @@
-PREHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT
-
-DROP TABLE Employee_Part
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT
-
-DROP TABLE Employee_Part
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE Employee_Part(employeeID int, employeeName String) partitioned by (employeeSalary double, country string)
-row format delimited fields terminated by '|' stored as textfile
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@Employee_Part
-POSTHOOK: query: CREATE TABLE Employee_Part(employeeID int, employeeName String) partitioned by (employeeSalary double, country string)
-row format delimited fields terminated by '|' stored as textfile
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@Employee_Part
-PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='USA')
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@employee_part
-POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='USA')
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@employee_part
-POSTHOOK: Output: default@employee_part@employeesalary=2000.0/country=USA
-PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='UK')
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@employee_part
-POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='UK')
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@employee_part
-POSTHOOK: Output: default@employee_part@employeesalary=2000.0/country=UK
-PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='USA')
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@employee_part
-POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='USA')
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@employee_part
-POSTHOOK: Output: default@employee_part@employeesalary=3000.0/country=USA
-PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='4000.0', country='USA')
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@employee_part
-POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='4000.0', country='USA')
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@employee_part
-POSTHOOK: Output: default@employee_part@employeesalary=4000.0/country=USA
-PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3500.0', country='UK')
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@employee_part
-POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3500.0', country='UK')
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@employee_part
-POSTHOOK: Output: default@employee_part@employeesalary=3500.0/country=UK
-PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='UK')
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@employee_part
-POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='UK')
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@employee_part
-POSTHOOK: Output: default@employee_part@employeesalary=3000.0/country=UK
-FAILED: SemanticException [Error 30007]: Invalid partitioning key/value specified in ANALYZE statement : {country=Canada, employeesalary=4000.0}
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.out b/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.out
new file mode 100644
index 0000000..3261f78
--- /dev/null
+++ b/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.out
@@ -0,0 +1,69 @@
+PREHOOK: query: DROP TABLE Employee_Part
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: DROP TABLE Employee_Part
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: CREATE TABLE Employee_Part(employeeID int, employeeName String) partitioned by (employeeSalary double, country string)
+row format delimited fields terminated by '|' stored as textfile
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@Employee_Part
+POSTHOOK: query: CREATE TABLE Employee_Part(employeeID int, employeeName String) partitioned by (employeeSalary double, country string)
+row format delimited fields terminated by '|' stored as textfile
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@Employee_Part
+PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='USA')
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@employee_part
+POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='USA')
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@employee_part
+POSTHOOK: Output: default@employee_part@employeesalary=2000.0/country=USA
+PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='UK')
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@employee_part
+POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='UK')
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@employee_part
+POSTHOOK: Output: default@employee_part@employeesalary=2000.0/country=UK
+PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='USA')
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@employee_part
+POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='USA')
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@employee_part
+POSTHOOK: Output: default@employee_part@employeesalary=3000.0/country=USA
+PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='4000.0', country='USA')
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@employee_part
+POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='4000.0', country='USA')
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@employee_part
+POSTHOOK: Output: default@employee_part@employeesalary=4000.0/country=USA
+PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3500.0', country='UK')
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@employee_part
+POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3500.0', country='UK')
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@employee_part
+POSTHOOK: Output: default@employee_part@employeesalary=3500.0/country=UK
+PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='UK')
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@employee_part
+POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='UK')
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@employee_part
+POSTHOOK: Output: default@employee_part@employeesalary=3000.0/country=UK
+FAILED: SemanticException [Error 30007]: Invalid partitioning key/value specified in ANALYZE statement : {employeesalary=4000.0, country=Canada}
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/results/clientpositive/authorization_explain.q.java1.7.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/authorization_explain.q.java1.7.out b/ql/src/test/results/clientpositive/authorization_explain.q.java1.7.out
deleted file mode 100644
index fefb50c..0000000
--- a/ql/src/test/results/clientpositive/authorization_explain.q.java1.7.out
+++ /dev/null
@@ -1,44 +0,0 @@
-Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product
-PREHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT
-
-explain authorization select * from src join srcpart
-PREHOOK: type: QUERY
-POSTHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT
-
-explain authorization select * from src join srcpart
-POSTHOOK: type: QUERY
-INPUTS:
- default@src
- default@srcpart
- default@srcpart@ds=2008-04-08/hr=11
- default@srcpart@ds=2008-04-08/hr=12
- default@srcpart@ds=2008-04-09/hr=11
- default@srcpart@ds=2008-04-09/hr=12
-OUTPUTS:
-#### A masked pattern was here ####
-CURRENT_USER:
- hive_test_user
-OPERATION:
- QUERY
-Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product
-PREHOOK: query: explain formatted authorization select * from src join srcpart
-PREHOOK: type: QUERY
-POSTHOOK: query: explain formatted authorization select * from src join srcpart
-POSTHOOK: type: QUERY
-#### A masked pattern was here ####
-PREHOOK: query: explain authorization use default
-PREHOOK: type: SWITCHDATABASE
-POSTHOOK: query: explain authorization use default
-POSTHOOK: type: SWITCHDATABASE
-INPUTS:
- database:default
-OUTPUTS:
-CURRENT_USER:
- hive_test_user
-OPERATION:
- SWITCHDATABASE
-PREHOOK: query: explain formatted authorization use default
-PREHOOK: type: SWITCHDATABASE
-POSTHOOK: query: explain formatted authorization use default
-POSTHOOK: type: SWITCHDATABASE
-{"INPUTS":["database:default"],"OUTPUTS":[],"CURRENT_USER":"hive_test_user","OPERATION":"SWITCHDATABASE"}
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/results/clientpositive/authorization_explain.q.java1.8.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/authorization_explain.q.java1.8.out b/ql/src/test/results/clientpositive/authorization_explain.q.java1.8.out
deleted file mode 100644
index b7ec209..0000000
--- a/ql/src/test/results/clientpositive/authorization_explain.q.java1.8.out
+++ /dev/null
@@ -1,47 +0,0 @@
-Warning: Shuffle Join JOIN[7][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product
-PREHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT
-
-explain authorization select * from src join srcpart
-PREHOOK: type: QUERY
-POSTHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT
-
-explain authorization select * from src join srcpart
-POSTHOOK: type: QUERY
-INPUTS:
- default@src
- default@srcpart
- default@srcpart@ds=2008-04-08/hr=11
- default@srcpart@ds=2008-04-08/hr=12
- default@srcpart@ds=2008-04-09/hr=11
- default@srcpart@ds=2008-04-09/hr=12
-OUTPUTS:
-#### A masked pattern was here ####
-CURRENT_USER:
- hive_test_user
-OPERATION:
- QUERY
-AUTHORIZATION_FAILURES:
- No privilege 'Select' found for inputs { database:default, table:src, columnName:key}
- No privilege 'Select' found for inputs { database:default, table:srcpart, columnName:key}
-Warning: Shuffle Join JOIN[7][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product
-PREHOOK: query: explain formatted authorization select * from src join srcpart
-PREHOOK: type: QUERY
-POSTHOOK: query: explain formatted authorization select * from src join srcpart
-POSTHOOK: type: QUERY
-#### A masked pattern was here ####
-PREHOOK: query: explain authorization use default
-PREHOOK: type: SWITCHDATABASE
-POSTHOOK: query: explain authorization use default
-POSTHOOK: type: SWITCHDATABASE
-INPUTS:
- database:default
-OUTPUTS:
-CURRENT_USER:
- hive_test_user
-OPERATION:
- SWITCHDATABASE
-PREHOOK: query: explain formatted authorization use default
-PREHOOK: type: SWITCHDATABASE
-POSTHOOK: query: explain formatted authorization use default
-POSTHOOK: type: SWITCHDATABASE
-{"INPUTS":["database:default"],"OUTPUTS":[],"CURRENT_USER":"hive_test_user","OPERATION":"SWITCHDATABASE"}
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/results/clientpositive/authorization_explain.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/authorization_explain.q.out b/ql/src/test/results/clientpositive/authorization_explain.q.out
new file mode 100644
index 0000000..851b845
--- /dev/null
+++ b/ql/src/test/results/clientpositive/authorization_explain.q.out
@@ -0,0 +1,40 @@
+Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product
+PREHOOK: query: explain authorization select * from src join srcpart
+PREHOOK: type: QUERY
+POSTHOOK: query: explain authorization select * from src join srcpart
+POSTHOOK: type: QUERY
+INPUTS:
+ default@src
+ default@srcpart
+ default@srcpart@ds=2008-04-08/hr=11
+ default@srcpart@ds=2008-04-08/hr=12
+ default@srcpart@ds=2008-04-09/hr=11
+ default@srcpart@ds=2008-04-09/hr=12
+OUTPUTS:
+#### A masked pattern was here ####
+CURRENT_USER:
+ hive_test_user
+OPERATION:
+ QUERY
+Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product
+PREHOOK: query: explain formatted authorization select * from src join srcpart
+PREHOOK: type: QUERY
+POSTHOOK: query: explain formatted authorization select * from src join srcpart
+POSTHOOK: type: QUERY
+#### A masked pattern was here ####
+PREHOOK: query: explain authorization use default
+PREHOOK: type: SWITCHDATABASE
+POSTHOOK: query: explain authorization use default
+POSTHOOK: type: SWITCHDATABASE
+INPUTS:
+ database:default
+OUTPUTS:
+CURRENT_USER:
+ hive_test_user
+OPERATION:
+ SWITCHDATABASE
+PREHOOK: query: explain formatted authorization use default
+PREHOOK: type: SWITCHDATABASE
+POSTHOOK: query: explain formatted authorization use default
+POSTHOOK: type: SWITCHDATABASE
+{"INPUTS":["database:default"],"OUTPUTS":[],"CURRENT_USER":"hive_test_user","OPERATION":"SWITCHDATABASE"}
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/results/clientpositive/avro_date.q.java1.7.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/avro_date.q.java1.7.out b/ql/src/test/results/clientpositive/avro_date.q.java1.7.out
deleted file mode 100644
index 501b983..0000000
--- a/ql/src/test/results/clientpositive/avro_date.q.java1.7.out
+++ /dev/null
@@ -1,130 +0,0 @@
-PREHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT
-
-DROP TABLE avro_date_staging
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT
-
-DROP TABLE avro_date_staging
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: DROP TABLE avro_date
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE avro_date
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: DROP TABLE avro_date_casts
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE avro_date_casts
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE avro_date_staging (d date, m1 map<string, date>, l1 array<date>)
- ROW FORMAT DELIMITED FIELDS TERMINATED BY '|'
- COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':'
- STORED AS TEXTFILE
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@avro_date_staging
-POSTHOOK: query: CREATE TABLE avro_date_staging (d date, m1 map<string, date>, l1 array<date>)
- ROW FORMAT DELIMITED FIELDS TERMINATED BY '|'
- COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':'
- STORED AS TEXTFILE
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@avro_date_staging
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/avro_date.txt' OVERWRITE INTO TABLE avro_date_staging
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@avro_date_staging
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/avro_date.txt' OVERWRITE INTO TABLE avro_date_staging
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@avro_date_staging
-PREHOOK: query: CREATE TABLE avro_date (d date, m1 map<string, date>, l1 array<date>)
- PARTITIONED BY (p1 int, p2 date)
- ROW FORMAT DELIMITED FIELDS TERMINATED BY '|'
- COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':'
- STORED AS AVRO
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@avro_date
-POSTHOOK: query: CREATE TABLE avro_date (d date, m1 map<string, date>, l1 array<date>)
- PARTITIONED BY (p1 int, p2 date)
- ROW FORMAT DELIMITED FIELDS TERMINATED BY '|'
- COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':'
- STORED AS AVRO
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@avro_date
-PREHOOK: query: INSERT OVERWRITE TABLE avro_date PARTITION(p1=2, p2='2014-09-26') SELECT * FROM avro_date_staging
-PREHOOK: type: QUERY
-PREHOOK: Input: default@avro_date_staging
-PREHOOK: Output: default@avro_date@p1=2/p2=2014-09-26
-POSTHOOK: query: INSERT OVERWRITE TABLE avro_date PARTITION(p1=2, p2='2014-09-26') SELECT * FROM avro_date_staging
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@avro_date_staging
-POSTHOOK: Output: default@avro_date@p1=2/p2=2014-09-26
-POSTHOOK: Lineage: avro_date PARTITION(p1=2,p2=2014-09-26).d SIMPLE [(avro_date_staging)avro_date_staging.FieldSchema(name:d, type:date, comment:null), ]
-POSTHOOK: Lineage: avro_date PARTITION(p1=2,p2=2014-09-26).l1 SIMPLE [(avro_date_staging)avro_date_staging.FieldSchema(name:l1, type:array<date>, comment:null), ]
-POSTHOOK: Lineage: avro_date PARTITION(p1=2,p2=2014-09-26).m1 SIMPLE [(avro_date_staging)avro_date_staging.FieldSchema(name:m1, type:map<string,date>, comment:null), ]
-PREHOOK: query: SELECT * FROM avro_date
-PREHOOK: type: QUERY
-PREHOOK: Input: default@avro_date
-PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT * FROM avro_date
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@avro_date
-POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
-#### A masked pattern was here ####
-2012-02-21 {"foo":"1980-12-16","bar":"1998-05-07"} ["2011-09-04","2011-09-05"] 2 2014-09-26
-2014-02-11 {"baz":"1981-12-16"} ["2011-09-05"] 2 2014-09-26
-1947-02-11 {"baz":"1921-12-16"} ["2011-09-05"] 2 2014-09-26
-8200-02-11 {"baz":"6981-12-16"} ["1039-09-05"] 2 2014-09-26
-PREHOOK: query: SELECT d, COUNT(d) FROM avro_date GROUP BY d
-PREHOOK: type: QUERY
-PREHOOK: Input: default@avro_date
-PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT d, COUNT(d) FROM avro_date GROUP BY d
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@avro_date
-POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
-#### A masked pattern was here ####
-1947-02-11 1
-2012-02-21 1
-2014-02-11 1
-8200-02-11 1
-PREHOOK: query: SELECT * FROM avro_date WHERE d!='1947-02-11'
-PREHOOK: type: QUERY
-PREHOOK: Input: default@avro_date
-PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT * FROM avro_date WHERE d!='1947-02-11'
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@avro_date
-POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
-#### A masked pattern was here ####
-2012-02-21 {"foo":"1980-12-16","bar":"1998-05-07"} ["2011-09-04","2011-09-05"] 2 2014-09-26
-2014-02-11 {"baz":"1981-12-16"} ["2011-09-05"] 2 2014-09-26
-8200-02-11 {"baz":"6981-12-16"} ["1039-09-05"] 2 2014-09-26
-PREHOOK: query: SELECT * FROM avro_date WHERE d<'2014-12-21'
-PREHOOK: type: QUERY
-PREHOOK: Input: default@avro_date
-PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT * FROM avro_date WHERE d<'2014-12-21'
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@avro_date
-POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
-#### A masked pattern was here ####
-2012-02-21 {"foo":"1980-12-16","bar":"1998-05-07"} ["2011-09-04","2011-09-05"] 2 2014-09-26
-2014-02-11 {"baz":"1981-12-16"} ["2011-09-05"] 2 2014-09-26
-1947-02-11 {"baz":"1921-12-16"} ["2011-09-05"] 2 2014-09-26
-PREHOOK: query: SELECT * FROM avro_date WHERE d>'8000-12-01'
-PREHOOK: type: QUERY
-PREHOOK: Input: default@avro_date
-PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT * FROM avro_date WHERE d>'8000-12-01'
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@avro_date
-POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
-#### A masked pattern was here ####
-8200-02-11 {"baz":"6981-12-16"} ["1039-09-05"] 2 2014-09-26
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/results/clientpositive/avro_date.q.java1.8.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/avro_date.q.java1.8.out b/ql/src/test/results/clientpositive/avro_date.q.java1.8.out
deleted file mode 100644
index dea51c6..0000000
--- a/ql/src/test/results/clientpositive/avro_date.q.java1.8.out
+++ /dev/null
@@ -1,130 +0,0 @@
-PREHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT
-
-DROP TABLE avro_date_staging
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT
-
-DROP TABLE avro_date_staging
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: DROP TABLE avro_date
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE avro_date
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: DROP TABLE avro_date_casts
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE avro_date_casts
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE avro_date_staging (d date, m1 map<string, date>, l1 array<date>)
- ROW FORMAT DELIMITED FIELDS TERMINATED BY '|'
- COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':'
- STORED AS TEXTFILE
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@avro_date_staging
-POSTHOOK: query: CREATE TABLE avro_date_staging (d date, m1 map<string, date>, l1 array<date>)
- ROW FORMAT DELIMITED FIELDS TERMINATED BY '|'
- COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':'
- STORED AS TEXTFILE
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@avro_date_staging
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/avro_date.txt' OVERWRITE INTO TABLE avro_date_staging
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@avro_date_staging
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/avro_date.txt' OVERWRITE INTO TABLE avro_date_staging
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@avro_date_staging
-PREHOOK: query: CREATE TABLE avro_date (d date, m1 map<string, date>, l1 array<date>)
- PARTITIONED BY (p1 int, p2 date)
- ROW FORMAT DELIMITED FIELDS TERMINATED BY '|'
- COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':'
- STORED AS AVRO
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@avro_date
-POSTHOOK: query: CREATE TABLE avro_date (d date, m1 map<string, date>, l1 array<date>)
- PARTITIONED BY (p1 int, p2 date)
- ROW FORMAT DELIMITED FIELDS TERMINATED BY '|'
- COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':'
- STORED AS AVRO
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@avro_date
-PREHOOK: query: INSERT OVERWRITE TABLE avro_date PARTITION(p1=2, p2='2014-09-26') SELECT * FROM avro_date_staging
-PREHOOK: type: QUERY
-PREHOOK: Input: default@avro_date_staging
-PREHOOK: Output: default@avro_date@p1=2/p2=2014-09-26
-POSTHOOK: query: INSERT OVERWRITE TABLE avro_date PARTITION(p1=2, p2='2014-09-26') SELECT * FROM avro_date_staging
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@avro_date_staging
-POSTHOOK: Output: default@avro_date@p1=2/p2=2014-09-26
-POSTHOOK: Lineage: avro_date PARTITION(p1=2,p2=2014-09-26).d SIMPLE [(avro_date_staging)avro_date_staging.FieldSchema(name:d, type:date, comment:null), ]
-POSTHOOK: Lineage: avro_date PARTITION(p1=2,p2=2014-09-26).l1 SIMPLE [(avro_date_staging)avro_date_staging.FieldSchema(name:l1, type:array<date>, comment:null), ]
-POSTHOOK: Lineage: avro_date PARTITION(p1=2,p2=2014-09-26).m1 SIMPLE [(avro_date_staging)avro_date_staging.FieldSchema(name:m1, type:map<string,date>, comment:null), ]
-PREHOOK: query: SELECT * FROM avro_date
-PREHOOK: type: QUERY
-PREHOOK: Input: default@avro_date
-PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT * FROM avro_date
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@avro_date
-POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
-#### A masked pattern was here ####
-2012-02-21 {"bar":"1998-05-07","foo":"1980-12-16"} ["2011-09-04","2011-09-05"] 2 2014-09-26
-2014-02-11 {"baz":"1981-12-16"} ["2011-09-05"] 2 2014-09-26
-1947-02-11 {"baz":"1921-12-16"} ["2011-09-05"] 2 2014-09-26
-8200-02-11 {"baz":"6981-12-16"} ["1039-09-05"] 2 2014-09-26
-PREHOOK: query: SELECT d, COUNT(d) FROM avro_date GROUP BY d
-PREHOOK: type: QUERY
-PREHOOK: Input: default@avro_date
-PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT d, COUNT(d) FROM avro_date GROUP BY d
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@avro_date
-POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
-#### A masked pattern was here ####
-1947-02-11 1
-2012-02-21 1
-2014-02-11 1
-8200-02-11 1
-PREHOOK: query: SELECT * FROM avro_date WHERE d!='1947-02-11'
-PREHOOK: type: QUERY
-PREHOOK: Input: default@avro_date
-PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT * FROM avro_date WHERE d!='1947-02-11'
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@avro_date
-POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
-#### A masked pattern was here ####
-2012-02-21 {"bar":"1998-05-07","foo":"1980-12-16"} ["2011-09-04","2011-09-05"] 2 2014-09-26
-2014-02-11 {"baz":"1981-12-16"} ["2011-09-05"] 2 2014-09-26
-8200-02-11 {"baz":"6981-12-16"} ["1039-09-05"] 2 2014-09-26
-PREHOOK: query: SELECT * FROM avro_date WHERE d<'2014-12-21'
-PREHOOK: type: QUERY
-PREHOOK: Input: default@avro_date
-PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT * FROM avro_date WHERE d<'2014-12-21'
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@avro_date
-POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
-#### A masked pattern was here ####
-2012-02-21 {"bar":"1998-05-07","foo":"1980-12-16"} ["2011-09-04","2011-09-05"] 2 2014-09-26
-2014-02-11 {"baz":"1981-12-16"} ["2011-09-05"] 2 2014-09-26
-1947-02-11 {"baz":"1921-12-16"} ["2011-09-05"] 2 2014-09-26
-PREHOOK: query: SELECT * FROM avro_date WHERE d>'8000-12-01'
-PREHOOK: type: QUERY
-PREHOOK: Input: default@avro_date
-PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT * FROM avro_date WHERE d>'8000-12-01'
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@avro_date
-POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
-#### A masked pattern was here ####
-8200-02-11 {"baz":"6981-12-16"} ["1039-09-05"] 2 2014-09-26
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/results/clientpositive/avro_date.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/avro_date.q.out b/ql/src/test/results/clientpositive/avro_date.q.out
new file mode 100644
index 0000000..32501cf
--- /dev/null
+++ b/ql/src/test/results/clientpositive/avro_date.q.out
@@ -0,0 +1,126 @@
+PREHOOK: query: DROP TABLE avro_date_staging
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: DROP TABLE avro_date_staging
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: DROP TABLE avro_date
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: DROP TABLE avro_date
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: DROP TABLE avro_date_casts
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: DROP TABLE avro_date_casts
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: CREATE TABLE avro_date_staging (d date, m1 map<string, date>, l1 array<date>)
+ ROW FORMAT DELIMITED FIELDS TERMINATED BY '|'
+ COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':'
+ STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@avro_date_staging
+POSTHOOK: query: CREATE TABLE avro_date_staging (d date, m1 map<string, date>, l1 array<date>)
+ ROW FORMAT DELIMITED FIELDS TERMINATED BY '|'
+ COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':'
+ STORED AS TEXTFILE
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@avro_date_staging
+PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/avro_date.txt' OVERWRITE INTO TABLE avro_date_staging
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@avro_date_staging
+POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/avro_date.txt' OVERWRITE INTO TABLE avro_date_staging
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@avro_date_staging
+PREHOOK: query: CREATE TABLE avro_date (d date, m1 map<string, date>, l1 array<date>)
+ PARTITIONED BY (p1 int, p2 date)
+ ROW FORMAT DELIMITED FIELDS TERMINATED BY '|'
+ COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':'
+ STORED AS AVRO
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@avro_date
+POSTHOOK: query: CREATE TABLE avro_date (d date, m1 map<string, date>, l1 array<date>)
+ PARTITIONED BY (p1 int, p2 date)
+ ROW FORMAT DELIMITED FIELDS TERMINATED BY '|'
+ COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':'
+ STORED AS AVRO
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@avro_date
+PREHOOK: query: INSERT OVERWRITE TABLE avro_date PARTITION(p1=2, p2='2014-09-26') SELECT * FROM avro_date_staging
+PREHOOK: type: QUERY
+PREHOOK: Input: default@avro_date_staging
+PREHOOK: Output: default@avro_date@p1=2/p2=2014-09-26
+POSTHOOK: query: INSERT OVERWRITE TABLE avro_date PARTITION(p1=2, p2='2014-09-26') SELECT * FROM avro_date_staging
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@avro_date_staging
+POSTHOOK: Output: default@avro_date@p1=2/p2=2014-09-26
+POSTHOOK: Lineage: avro_date PARTITION(p1=2,p2=2014-09-26).d SIMPLE [(avro_date_staging)avro_date_staging.FieldSchema(name:d, type:date, comment:null), ]
+POSTHOOK: Lineage: avro_date PARTITION(p1=2,p2=2014-09-26).l1 SIMPLE [(avro_date_staging)avro_date_staging.FieldSchema(name:l1, type:array<date>, comment:null), ]
+POSTHOOK: Lineage: avro_date PARTITION(p1=2,p2=2014-09-26).m1 SIMPLE [(avro_date_staging)avro_date_staging.FieldSchema(name:m1, type:map<string,date>, comment:null), ]
+PREHOOK: query: SELECT * FROM avro_date
+PREHOOK: type: QUERY
+PREHOOK: Input: default@avro_date
+PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
+#### A masked pattern was here ####
+POSTHOOK: query: SELECT * FROM avro_date
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@avro_date
+POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
+#### A masked pattern was here ####
+2012-02-21 {"bar":"1998-05-07","foo":"1980-12-16"} ["2011-09-04","2011-09-05"] 2 2014-09-26
+2014-02-11 {"baz":"1981-12-16"} ["2011-09-05"] 2 2014-09-26
+1947-02-11 {"baz":"1921-12-16"} ["2011-09-05"] 2 2014-09-26
+8200-02-11 {"baz":"6981-12-16"} ["1039-09-05"] 2 2014-09-26
+PREHOOK: query: SELECT d, COUNT(d) FROM avro_date GROUP BY d
+PREHOOK: type: QUERY
+PREHOOK: Input: default@avro_date
+PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
+#### A masked pattern was here ####
+POSTHOOK: query: SELECT d, COUNT(d) FROM avro_date GROUP BY d
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@avro_date
+POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
+#### A masked pattern was here ####
+1947-02-11 1
+2012-02-21 1
+2014-02-11 1
+8200-02-11 1
+PREHOOK: query: SELECT * FROM avro_date WHERE d!='1947-02-11'
+PREHOOK: type: QUERY
+PREHOOK: Input: default@avro_date
+PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
+#### A masked pattern was here ####
+POSTHOOK: query: SELECT * FROM avro_date WHERE d!='1947-02-11'
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@avro_date
+POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
+#### A masked pattern was here ####
+2012-02-21 {"bar":"1998-05-07","foo":"1980-12-16"} ["2011-09-04","2011-09-05"] 2 2014-09-26
+2014-02-11 {"baz":"1981-12-16"} ["2011-09-05"] 2 2014-09-26
+8200-02-11 {"baz":"6981-12-16"} ["1039-09-05"] 2 2014-09-26
+PREHOOK: query: SELECT * FROM avro_date WHERE d<'2014-12-21'
+PREHOOK: type: QUERY
+PREHOOK: Input: default@avro_date
+PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
+#### A masked pattern was here ####
+POSTHOOK: query: SELECT * FROM avro_date WHERE d<'2014-12-21'
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@avro_date
+POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
+#### A masked pattern was here ####
+2012-02-21 {"bar":"1998-05-07","foo":"1980-12-16"} ["2011-09-04","2011-09-05"] 2 2014-09-26
+2014-02-11 {"baz":"1981-12-16"} ["2011-09-05"] 2 2014-09-26
+1947-02-11 {"baz":"1921-12-16"} ["2011-09-05"] 2 2014-09-26
+PREHOOK: query: SELECT * FROM avro_date WHERE d>'8000-12-01'
+PREHOOK: type: QUERY
+PREHOOK: Input: default@avro_date
+PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
+#### A masked pattern was here ####
+POSTHOOK: query: SELECT * FROM avro_date WHERE d>'8000-12-01'
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@avro_date
+POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26
+#### A masked pattern was here ####
+8200-02-11 {"baz":"6981-12-16"} ["1039-09-05"] 2 2014-09-26
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.java1.7.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.java1.7.out b/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.java1.7.out
deleted file mode 100644
index 8f8065e..0000000
--- a/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.java1.7.out
+++ /dev/null
@@ -1,57 +0,0 @@
-PREHOOK: query: -- These test attempts to deserialize an Avro file that contains map null values, and the file schema
--- vs record schema have the null values in different positions
--- i.e.
--- fileSchema = [{ "type" : "map", "values" : ["string","null"]}, "null"]
--- recordSchema = ["null", { "type" : "map", "values" : ["string","null"]}]
-
--- JAVA_VERSION_SPECIFIC_OUTPUT
-
-DROP TABLE IF EXISTS avro_table
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: -- These test attempts to deserialize an Avro file that contains map null values, and the file schema
--- vs record schema have the null values in different positions
--- i.e.
--- fileSchema = [{ "type" : "map", "values" : ["string","null"]}, "null"]
--- recordSchema = ["null", { "type" : "map", "values" : ["string","null"]}]
-
--- JAVA_VERSION_SPECIFIC_OUTPUT
-
-DROP TABLE IF EXISTS avro_table
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE avro_table (avreau_col_1 map<string,string>) STORED AS AVRO
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@avro_table
-POSTHOOK: query: CREATE TABLE avro_table (avreau_col_1 map<string,string>) STORED AS AVRO
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@avro_table
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/map_null_val.avro' OVERWRITE INTO TABLE avro_table
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@avro_table
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/map_null_val.avro' OVERWRITE INTO TABLE avro_table
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@avro_table
-PREHOOK: query: SELECT * FROM avro_table
-PREHOOK: type: QUERY
-PREHOOK: Input: default@avro_table
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT * FROM avro_table
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@avro_table
-#### A masked pattern was here ####
-{"key4":null,"key3":"val3"}
-{"key4":null,"key3":"val3"}
-{"key2":"val2","key1":null}
-{"key4":null,"key3":"val3"}
-{"key4":null,"key3":"val3"}
-PREHOOK: query: DROP TABLE avro_table
-PREHOOK: type: DROPTABLE
-PREHOOK: Input: default@avro_table
-PREHOOK: Output: default@avro_table
-POSTHOOK: query: DROP TABLE avro_table
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Input: default@avro_table
-POSTHOOK: Output: default@avro_table
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.java1.8.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.java1.8.out b/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.java1.8.out
deleted file mode 100644
index 127d8b3..0000000
--- a/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.java1.8.out
+++ /dev/null
@@ -1,57 +0,0 @@
-PREHOOK: query: -- These test attempts to deserialize an Avro file that contains map null values, and the file schema
--- vs record schema have the null values in different positions
--- i.e.
--- fileSchema = [{ "type" : "map", "values" : ["string","null"]}, "null"]
--- recordSchema = ["null", { "type" : "map", "values" : ["string","null"]}]
-
--- JAVA_VERSION_SPECIFIC_OUTPUT
-
-DROP TABLE IF EXISTS avro_table
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: -- These test attempts to deserialize an Avro file that contains map null values, and the file schema
--- vs record schema have the null values in different positions
--- i.e.
--- fileSchema = [{ "type" : "map", "values" : ["string","null"]}, "null"]
--- recordSchema = ["null", { "type" : "map", "values" : ["string","null"]}]
-
--- JAVA_VERSION_SPECIFIC_OUTPUT
-
-DROP TABLE IF EXISTS avro_table
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE avro_table (avreau_col_1 map<string,string>) STORED AS AVRO
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@avro_table
-POSTHOOK: query: CREATE TABLE avro_table (avreau_col_1 map<string,string>) STORED AS AVRO
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@avro_table
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/map_null_val.avro' OVERWRITE INTO TABLE avro_table
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@avro_table
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/map_null_val.avro' OVERWRITE INTO TABLE avro_table
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@avro_table
-PREHOOK: query: SELECT * FROM avro_table
-PREHOOK: type: QUERY
-PREHOOK: Input: default@avro_table
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT * FROM avro_table
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@avro_table
-#### A masked pattern was here ####
-{"key3":"val3","key4":null}
-{"key3":"val3","key4":null}
-{"key1":null,"key2":"val2"}
-{"key3":"val3","key4":null}
-{"key3":"val3","key4":null}
-PREHOOK: query: DROP TABLE avro_table
-PREHOOK: type: DROPTABLE
-PREHOOK: Input: default@avro_table
-PREHOOK: Output: default@avro_table
-POSTHOOK: query: DROP TABLE avro_table
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Input: default@avro_table
-POSTHOOK: Output: default@avro_table
http://git-wip-us.apache.org/repos/asf/hive/blob/9349b8e5/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.out b/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.out
new file mode 100644
index 0000000..2d983f1
--- /dev/null
+++ b/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.out
@@ -0,0 +1,55 @@
+PREHOOK: query: -- These test attempts to deserialize an Avro file that contains map null values, and the file schema
+-- vs record schema have the null values in different positions
+-- i.e.
+-- fileSchema = [{ "type" : "map", "values" : ["string","null"]}, "null"]
+-- recordSchema = ["null", { "type" : "map", "values" : ["string","null"]}]
+
+
+DROP TABLE IF EXISTS avro_table
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: -- These test attempts to deserialize an Avro file that contains map null values, and the file schema
+-- vs record schema have the null values in different positions
+-- i.e.
+-- fileSchema = [{ "type" : "map", "values" : ["string","null"]}, "null"]
+-- recordSchema = ["null", { "type" : "map", "values" : ["string","null"]}]
+
+
+DROP TABLE IF EXISTS avro_table
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: CREATE TABLE avro_table (avreau_col_1 map<string,string>) STORED AS AVRO
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@avro_table
+POSTHOOK: query: CREATE TABLE avro_table (avreau_col_1 map<string,string>) STORED AS AVRO
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@avro_table
+PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/map_null_val.avro' OVERWRITE INTO TABLE avro_table
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@avro_table
+POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/map_null_val.avro' OVERWRITE INTO TABLE avro_table
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@avro_table
+PREHOOK: query: SELECT * FROM avro_table
+PREHOOK: type: QUERY
+PREHOOK: Input: default@avro_table
+#### A masked pattern was here ####
+POSTHOOK: query: SELECT * FROM avro_table
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@avro_table
+#### A masked pattern was here ####
+{"key3":"val3","key4":null}
+{"key3":"val3","key4":null}
+{"key1":null,"key2":"val2"}
+{"key3":"val3","key4":null}
+{"key3":"val3","key4":null}
+PREHOOK: query: DROP TABLE avro_table
+PREHOOK: type: DROPTABLE
+PREHOOK: Input: default@avro_table
+PREHOOK: Output: default@avro_table
+POSTHOOK: query: DROP TABLE avro_table
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Input: default@avro_table
+POSTHOOK: Output: default@avro_table