You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by ha...@apache.org on 2020/05/17 20:21:52 UTC
[hive] branch master updated: HIVE-23292 : Reduce PartitionDesc
payload in MapWork (Rajesh Balamohan, Ramesh Kumar,
Ashutosh Chauhan via Ashutosh Chauhan)
This is an automated email from the ASF dual-hosted git repository.
hashutosh pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hive.git
The following commit(s) were added to refs/heads/master by this push:
new b9b54c2 HIVE-23292 : Reduce PartitionDesc payload in MapWork (Rajesh Balamohan, Ramesh Kumar, Ashutosh Chauhan via Ashutosh Chauhan)
b9b54c2 is described below
commit b9b54c241cc46b1149e7dfc7ace7d1040b5e6787
Author: Ashutosh Chauhan <ha...@apache.org>
AuthorDate: Fri May 15 23:02:35 2020 -0700
HIVE-23292 : Reduce PartitionDesc payload in MapWork (Rajesh Balamohan, Ramesh Kumar, Ashutosh Chauhan via Ashutosh Chauhan)
---
.../insert_into_dynamic_partitions.q.out | 8 -
.../results/clientpositive/insert_into_table.q.out | 76 ---
.../insert_overwrite_directory.q.out | 32 --
.../insert_overwrite_dynamic_partitions.q.out | 8 -
.../clientpositive/insert_overwrite_table.q.out | 76 ---
.../write_final_output_blobstore.q.out | 64 ---
.../apache/hadoop/hive/ql/plan/PartitionDesc.java | 7 +-
.../org/apache/hadoop/hive/ql/plan/TableDesc.java | 2 -
.../clientpositive/beeline/smb_mapjoin_11.q.out | 34 --
.../clientpositive/beeline/smb_mapjoin_12.q.out | 28 -
.../clientpositive/beeline/smb_mapjoin_13.q.out | 30 --
.../clientpositive/binary_output_format.q.out | 80 ---
.../results/clientpositive/bucket_map_join_1.q.out | 13 -
.../results/clientpositive/bucket_map_join_2.q.out | 13 -
.../clientpositive/bucket_map_join_spark1.q.out | 86 ---
.../clientpositive/bucket_map_join_spark2.q.out | 86 ---
.../clientpositive/bucket_map_join_spark3.q.out | 86 ---
.../clientpositive/bucket_map_join_spark4.q.out | 90 ----
.../results/clientpositive/bucketcontext_1.q.out | 45 --
.../results/clientpositive/bucketcontext_2.q.out | 45 --
.../results/clientpositive/bucketcontext_3.q.out | 36 --
.../results/clientpositive/bucketcontext_4.q.out | 36 --
.../results/clientpositive/bucketcontext_5.q.out | 26 -
.../results/clientpositive/bucketcontext_6.q.out | 36 --
.../results/clientpositive/bucketcontext_7.q.out | 54 --
.../results/clientpositive/bucketcontext_8.q.out | 54 --
.../results/clientpositive/bucketmapjoin10.q.out | 36 --
.../results/clientpositive/bucketmapjoin11.q.out | 72 ---
.../results/clientpositive/bucketmapjoin12.q.out | 38 --
.../results/clientpositive/bucketmapjoin13.q.out | 90 ----
.../results/clientpositive/bucketmapjoin5.q.out | 164 ------
.../results/clientpositive/bucketmapjoin8.q.out | 36 --
.../results/clientpositive/bucketmapjoin9.q.out | 36 --
.../clientpositive/bucketmapjoin_negative.q.out | 86 ---
.../clientpositive/bucketmapjoin_negative2.q.out | 95 ----
.../clientpositive/bucketmapjoin_negative3.q.out | 117 ----
.../druid/druidkafkamini_basic.q.out | 32 --
.../druid/druidmini_expressions.q.out | 16 -
.../encryption_join_unencrypted_tbl.q.out | 32 --
...ption_join_with_different_encryption_keys.q.out | 32 --
.../erasurecoding/erasure_explain.q.out | 48 --
.../test/results/clientpositive/join_map_ppr.q.out | 152 ------
.../kafka/kafka_storage_handler.q.out | 28 -
.../results/clientpositive/list_bucket_dml_8.q.out | 53 --
.../clientpositive/llap/acid_bucket_pruning.q.out | 15 -
.../clientpositive/llap/acid_nullscan.q.out | 15 -
.../llap/alter_partition_coltype.q.out | 48 --
.../llap/analyze_table_null_partition.q.out | 48 --
.../clientpositive/llap/autoColumnStats_5a.q.out | 20 -
.../clientpositive/llap/autoColumnStats_8.q.out | 60 ---
.../llap/auto_join_reordering_values.q.out | 80 ---
.../llap/auto_sortmerge_join_1.q.out | 81 ---
.../llap/auto_sortmerge_join_11.q.out | 126 -----
.../llap/auto_sortmerge_join_12.q.out | 45 --
.../llap/auto_sortmerge_join_2.q.out | 54 --
.../llap/auto_sortmerge_join_3.q.out | 81 ---
.../llap/auto_sortmerge_join_4.q.out | 81 ---
.../llap/auto_sortmerge_join_5.q.out | 78 ---
.../llap/auto_sortmerge_join_7.q.out | 108 ----
.../llap/auto_sortmerge_join_8.q.out | 108 ----
.../test/results/clientpositive/llap/bucket1.q.out | 30 --
.../test/results/clientpositive/llap/bucket2.q.out | 30 --
.../test/results/clientpositive/llap/bucket3.q.out | 20 -
.../test/results/clientpositive/llap/bucket4.q.out | 30 --
.../test/results/clientpositive/llap/bucket5.q.out | 94 ----
.../results/clientpositive/llap/bucket_many.q.out | 30 --
.../clientpositive/llap/bucket_map_join_tez2.q.out | 16 -
.../clientpositive/llap/bucket_num_reducers.q.out | 30 --
.../clientpositive/llap/bucket_num_reducers2.q.out | 30 --
.../clientpositive/llap/bucketmapjoin1.q.out | 76 ---
.../clientpositive/llap/bucketmapjoin2.q.out | 111 ----
.../clientpositive/llap/bucketmapjoin3.q.out | 68 ---
.../clientpositive/llap/bucketmapjoin4.q.out | 84 ---
.../clientpositive/llap/bucketmapjoin7.q.out | 18 -
.../llap/cbo_rp_outer_join_ppr.q.out | 80 ---
.../clientpositive/llap/cbo_stats_estimation.q.out | 28 -
.../clientpositive/llap/column_table_stats.q.out | 69 ---
.../llap/column_table_stats_orc.q.out | 49 --
.../clientpositive/llap/columnstats_partlvl.q.out | 22 -
.../clientpositive/llap/columnstats_tbllvl.q.out | 28 -
.../results/clientpositive/llap/comments.q.out | 32 --
.../llap/constantPropagateForSubQuery.q.out | 32 --
.../llap/disable_merge_for_bucketing.q.out | 30 --
.../llap/display_colstats_tbllvl.q.out | 14 -
.../llap/dynamic_partition_skip_default.q.out | 48 --
.../llap/dynamic_semijoin_reduction.q.out | 44 --
.../llap/dynamic_semijoin_user_level.q.out | 44 --
.../llap/extrapolate_part_stats_date.q.out | 48 --
.../llap/extrapolate_part_stats_full.q.out | 144 -----
.../llap/extrapolate_part_stats_partial.q.out | 456 ----------------
.../llap/extrapolate_part_stats_partial_ndv.q.out | 228 --------
.../results/clientpositive/llap/filter_aggr.q.out | 16 -
.../llap/filter_join_breaktask.q.out | 36 --
.../results/clientpositive/llap/filter_union.q.out | 32 --
.../clientpositive/llap/groupby_map_ppr.q.out | 40 --
.../llap/groupby_map_ppr_multi_distinct.q.out | 40 --
.../results/clientpositive/llap/groupby_ppr.q.out | 40 --
.../llap/groupby_ppr_multi_distinct.q.out | 80 ---
.../clientpositive/llap/groupby_sort_1_23.q.out | 587 ---------------------
.../clientpositive/llap/groupby_sort_6.q.out | 59 ---
.../llap/groupby_sort_skew_1_23.q.out | 587 ---------------------
.../llap/infer_bucket_sort_num_buckets.q.out | 30 --
.../test/results/clientpositive/llap/input23.q.out | 12 -
.../test/results/clientpositive/llap/input42.q.out | 72 ---
.../results/clientpositive/llap/input_part1.q.out | 28 -
.../results/clientpositive/llap/input_part2.q.out | 56 --
.../results/clientpositive/llap/input_part7.q.out | 48 --
.../results/clientpositive/llap/input_part9.q.out | 24 -
.../test/results/clientpositive/llap/join17.q.out | 48 --
.../test/results/clientpositive/llap/join26.q.out | 60 ---
.../test/results/clientpositive/llap/join32.q.out | 60 ---
.../clientpositive/llap/join32_lessSize.q.out | 260 ---------
.../test/results/clientpositive/llap/join33.q.out | 60 ---
.../test/results/clientpositive/llap/join34.q.out | 72 ---
.../test/results/clientpositive/llap/join35.q.out | 72 ---
.../test/results/clientpositive/llap/join9.q.out | 44 --
.../clientpositive/llap/join_filters_overlap.q.out | 160 ------
.../clientpositive/llap/list_bucket_dml_1.q.out | 42 --
.../clientpositive/llap/list_bucket_dml_10.q.out | 22 -
.../clientpositive/llap/list_bucket_dml_11.q.out | 34 --
.../clientpositive/llap/list_bucket_dml_12.q.out | 46 --
.../clientpositive/llap/list_bucket_dml_13.q.out | 34 --
.../clientpositive/llap/list_bucket_dml_14.q.out | 32 --
.../clientpositive/llap/list_bucket_dml_2.q.out | 42 --
.../clientpositive/llap/list_bucket_dml_3.q.out | 42 --
.../clientpositive/llap/list_bucket_dml_4.q.out | 72 ---
.../clientpositive/llap/list_bucket_dml_5.q.out | 54 --
.../clientpositive/llap/list_bucket_dml_6.q.out | 84 ---
.../clientpositive/llap/list_bucket_dml_7.q.out | 84 ---
.../clientpositive/llap/list_bucket_dml_9.q.out | 72 ---
.../llap/list_bucket_query_multiskew_1.q.out | 48 --
.../llap/list_bucket_query_multiskew_2.q.out | 36 --
.../llap/list_bucket_query_multiskew_3.q.out | 36 --
.../llap/list_bucket_query_oneskew_1.q.out | 27 -
.../llap/list_bucket_query_oneskew_2.q.out | 36 --
.../llap/list_bucket_query_oneskew_3.q.out | 9 -
.../clientpositive/llap/llap_nullscan.q.out | 16 -
.../clientpositive/llap/load_dyn_part8.q.out | 60 ---
.../clientpositive/llap/louter_join_ppr.q.out | 160 ------
.../clientpositive/llap/mapjoin_mapjoin.q.out | 80 ---
.../test/results/clientpositive/llap/merge3.q.out | 76 ---
.../clientpositive/llap/metadataonly1.q.out | 240 ---------
.../llap/murmur_hash_migration.q.out | 68 ---
.../llap/murmur_hash_migration2.q.out | 8 -
.../llap/offset_limit_global_optimizer.q.out | 384 --------------
.../clientpositive/llap/optimize_nullscan.q.out | 272 ----------
.../clientpositive/llap/outer_join_ppr.q.out | 80 ---
.../llap/parquet_vectorization_0.q.out | 80 ---
.../clientpositive/llap/partition_pruning.q.out | 60 ---
ql/src/test/results/clientpositive/llap/pcr.q.out | 568 --------------------
ql/src/test/results/clientpositive/llap/pcs.q.out | 252 ---------
.../results/clientpositive/llap/pointlookup2.q.out | 328 ------------
.../results/clientpositive/llap/pointlookup3.q.out | 288 ----------
.../results/clientpositive/llap/pointlookup4.q.out | 48 --
.../clientpositive/llap/ppd_join_filter.q.out | 64 ---
.../clientpositive/llap/ppd_union_view.q.out | 36 --
.../test/results/clientpositive/llap/ppd_vc.q.out | 112 ----
.../clientpositive/llap/ppr_allchildsarenull.q.out | 72 ---
.../test/results/clientpositive/llap/push_or.q.out | 24 -
.../llap/rand_partitionpruner2.q.out | 40 --
.../llap/rand_partitionpruner3.q.out | 24 -
.../clientpositive/llap/reduce_deduplicate.q.out | 36 --
.../clientpositive/llap/router_join_ppr.q.out | 160 ------
.../test/results/clientpositive/llap/sample1.q.out | 28 -
.../results/clientpositive/llap/sample10.q.out | 40 --
.../test/results/clientpositive/llap/sample5.q.out | 31 --
.../test/results/clientpositive/llap/sample6.q.out | 136 -----
.../test/results/clientpositive/llap/sample7.q.out | 31 --
.../test/results/clientpositive/llap/sample8.q.out | 60 ---
.../results/clientpositive/llap/sharedwork.q.out | 112 ----
.../clientpositive/llap/smb_mapjoin_15.q.out | 120 -----
.../test/results/clientpositive/llap/stats0.q.out | 64 ---
.../test/results/clientpositive/llap/stats11.q.out | 76 ---
.../test/results/clientpositive/llap/stats12.q.out | 18 -
.../test/results/clientpositive/llap/stats13.q.out | 9 -
.../test/results/clientpositive/llap/stats3.q.out | 8 -
.../llap/temp_table_alter_partition_coltype.q.out | 120 -----
.../llap/temp_table_display_colstats_tbllvl.q.out | 12 -
.../llap/tez_fixed_bucket_pruning.q.out | 156 ------
.../llap/tez_join_result_complex.q.out | 56 --
.../clientpositive/llap/topnkey_windowing.q.out | 32 --
.../clientpositive/llap/unionDistinct_1.q.out | 234 --------
.../clientpositive/llap/vectorization_0.q.out | 80 ---
.../results/clientpositive/regexp_extract.q.out | 32 --
.../clientpositive/serde_user_properties.q.out | 48 --
.../clientpositive/sort_merge_join_desc_5.q.out | 10 -
.../clientpositive/sort_merge_join_desc_6.q.out | 20 -
.../clientpositive/sort_merge_join_desc_7.q.out | 40 --
.../temp_table_partition_pruning.q.out | 90 ----
ql/src/test/results/clientpositive/timestamp.q.out | 16 -
.../results/clientpositive/transform_ppr1.q.out | 48 --
.../results/clientpositive/transform_ppr2.q.out | 24 -
.../truncate_column_list_bucket.q.out | 22 -
.../test/results/clientpositive/udf_explode.q.out | 32 --
.../test/results/clientpositive/udtf_explode.q.out | 32 --
ql/src/test/results/clientpositive/union22.q.out | 81 ---
ql/src/test/results/clientpositive/union24.q.out | 192 -------
ql/src/test/results/clientpositive/union_ppr.q.out | 24 -
.../hive/serde2/dynamic_type/DynamicSerDe.java | 83 +++
.../hive/metastore/utils/MetaStoreUtils.java | 22 +-
200 files changed, 98 insertions(+), 14677 deletions(-)
diff --git a/itests/hive-blobstore/src/test/results/clientpositive/insert_into_dynamic_partitions.q.out b/itests/hive-blobstore/src/test/results/clientpositive/insert_into_dynamic_partitions.q.out
index 3783c15..fd82a0c 100644
--- a/itests/hive-blobstore/src/test/results/clientpositive/insert_into_dynamic_partitions.q.out
+++ b/itests/hive-blobstore/src/test/results/clientpositive/insert_into_dynamic_partitions.q.out
@@ -129,11 +129,9 @@ STAGE PLANS:
bucketing_version 2
column.name.delimiter ,
columns
- columns.comments
columns.types
#### A masked pattern was here ####
name _dummy_database._dummy_table
- serialization.ddl struct _dummy_table { }
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -141,7 +139,6 @@ STAGE PLANS:
input format: org.apache.hadoop.hive.ql.io.NullRowsInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns
@@ -149,7 +146,6 @@ STAGE PLANS:
columns.types
#### A masked pattern was here ####
name _dummy_database._dummy_table
- serialization.ddl struct _dummy_table { }
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -187,10 +183,8 @@ STAGE PLANS:
name default.table1
partition_columns key
partition_columns.types string
- serialization.ddl struct table1 { i32 id}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
TotalFiles: 1
@@ -220,10 +214,8 @@ STAGE PLANS:
name default.table1
partition_columns key
partition_columns.types string
- serialization.ddl struct table1 { i32 id}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
diff --git a/itests/hive-blobstore/src/test/results/clientpositive/insert_into_table.q.out b/itests/hive-blobstore/src/test/results/clientpositive/insert_into_table.q.out
index 92c785c..ce071cb 100644
--- a/itests/hive-blobstore/src/test/results/clientpositive/insert_into_table.q.out
+++ b/itests/hive-blobstore/src/test/results/clientpositive/insert_into_table.q.out
@@ -90,8 +90,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns id
@@ -100,14 +98,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location ### test.blobstore.path ###/table1
name default.table1
- numFiles 2
- numRows 2
- rawDataSize 2
- serialization.ddl struct table1 { i32 id}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
TotalFiles: 1
@@ -145,11 +137,9 @@ STAGE PLANS:
bucketing_version 2
column.name.delimiter ,
columns
- columns.comments
columns.types
#### A masked pattern was here ####
name _dummy_database._dummy_table
- serialization.ddl struct _dummy_table { }
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -157,7 +147,6 @@ STAGE PLANS:
input format: org.apache.hadoop.hive.ql.io.NullRowsInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns
@@ -165,7 +154,6 @@ STAGE PLANS:
columns.types
#### A masked pattern was here ####
name _dummy_database._dummy_table
- serialization.ddl struct _dummy_table { }
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -217,8 +205,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns id
@@ -227,14 +213,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location ### test.blobstore.path ###/table1
name default.table1
- numFiles 2
- numRows 2
- rawDataSize 2
- serialization.ddl struct table1 { i32 id}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
@@ -263,8 +243,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns id
@@ -273,14 +251,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location ### test.blobstore.path ###/table1
name default.table1
- numFiles 2
- numRows 2
- rawDataSize 2
- serialization.ddl struct table1 { i32 id}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
TotalFiles: 1
@@ -295,8 +267,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns id
@@ -305,21 +275,13 @@ STAGE PLANS:
#### A masked pattern was here ####
location ### test.blobstore.path ###/table1
name default.table1
- numFiles 2
- numRows 2
- rawDataSize 2
- serialization.ddl struct table1 { i32 id}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns id
@@ -328,14 +290,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location ### test.blobstore.path ###/table1
name default.table1
- numFiles 2
- numRows 2
- rawDataSize 2
- serialization.ddl struct table1 { i32 id}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
name: default.table1
@@ -351,8 +307,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns id
@@ -361,14 +315,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location ### test.blobstore.path ###/table1
name default.table1
- numFiles 2
- numRows 2
- rawDataSize 2
- serialization.ddl struct table1 { i32 id}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
@@ -387,8 +335,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns id
@@ -397,14 +343,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location ### test.blobstore.path ###/table1
name default.table1
- numFiles 2
- numRows 2
- rawDataSize 2
- serialization.ddl struct table1 { i32 id}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
TotalFiles: 1
@@ -419,8 +359,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns id
@@ -429,21 +367,13 @@ STAGE PLANS:
#### A masked pattern was here ####
location ### test.blobstore.path ###/table1
name default.table1
- numFiles 2
- numRows 2
- rawDataSize 2
- serialization.ddl struct table1 { i32 id}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns id
@@ -452,14 +382,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location ### test.blobstore.path ###/table1
name default.table1
- numFiles 2
- numRows 2
- rawDataSize 2
- serialization.ddl struct table1 { i32 id}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
name: default.table1
diff --git a/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_directory.q.out b/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_directory.q.out
index a113a22..2eec9f2 100644
--- a/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_directory.q.out
+++ b/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_directory.q.out
@@ -120,31 +120,21 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true","key":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns id,key
- columns.comments
columns.types int:string
field.delim ,
#### A masked pattern was here ####
name default.table1
- numFiles 2
- numRows 2
- rawDataSize 8
- serialization.ddl struct table1 { i32 id, string key}
serialization.format ,
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 10
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true","key":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns id,key
@@ -153,14 +143,8 @@ STAGE PLANS:
field.delim ,
#### A masked pattern was here ####
name default.table1
- numFiles 2
- numRows 2
- rawDataSize 8
- serialization.ddl struct table1 { i32 id, string key}
serialization.format ,
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 10
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
name: default.table1
@@ -392,31 +376,21 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true","key":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns id,key
- columns.comments
columns.types int:string
field.delim ,
#### A masked pattern was here ####
name default.table1
- numFiles 2
- numRows 2
- rawDataSize 8
- serialization.ddl struct table1 { i32 id, string key}
serialization.format ,
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 10
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true","key":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns id,key
@@ -425,14 +399,8 @@ STAGE PLANS:
field.delim ,
#### A masked pattern was here ####
name default.table1
- numFiles 2
- numRows 2
- rawDataSize 8
- serialization.ddl struct table1 { i32 id, string key}
serialization.format ,
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 10
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
name: default.table1
diff --git a/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_dynamic_partitions.q.out b/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_dynamic_partitions.q.out
index 91e95c4..bdd82d4 100644
--- a/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_dynamic_partitions.q.out
+++ b/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_dynamic_partitions.q.out
@@ -147,11 +147,9 @@ STAGE PLANS:
bucketing_version 2
column.name.delimiter ,
columns
- columns.comments
columns.types
#### A masked pattern was here ####
name _dummy_database._dummy_table
- serialization.ddl struct _dummy_table { }
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -159,7 +157,6 @@ STAGE PLANS:
input format: org.apache.hadoop.hive.ql.io.NullRowsInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns
@@ -167,7 +164,6 @@ STAGE PLANS:
columns.types
#### A masked pattern was here ####
name _dummy_database._dummy_table
- serialization.ddl struct _dummy_table { }
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -205,10 +201,8 @@ STAGE PLANS:
name default.table1
partition_columns key
partition_columns.types string
- serialization.ddl struct table1 { i32 id}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
TotalFiles: 1
@@ -238,10 +232,8 @@ STAGE PLANS:
name default.table1
partition_columns key
partition_columns.types string
- serialization.ddl struct table1 { i32 id}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
diff --git a/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_table.q.out b/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_table.q.out
index 96e77ed..f65bf22 100644
--- a/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_table.q.out
+++ b/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_table.q.out
@@ -98,8 +98,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns id
@@ -108,14 +106,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location ### test.blobstore.path ###/table1
name default.table1
- numFiles 1
- numRows 1
- rawDataSize 1
- serialization.ddl struct table1 { i32 id}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
TotalFiles: 1
@@ -153,11 +145,9 @@ STAGE PLANS:
bucketing_version 2
column.name.delimiter ,
columns
- columns.comments
columns.types
#### A masked pattern was here ####
name _dummy_database._dummy_table
- serialization.ddl struct _dummy_table { }
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -165,7 +155,6 @@ STAGE PLANS:
input format: org.apache.hadoop.hive.ql.io.NullRowsInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns
@@ -173,7 +162,6 @@ STAGE PLANS:
columns.types
#### A masked pattern was here ####
name _dummy_database._dummy_table
- serialization.ddl struct _dummy_table { }
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -225,8 +213,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns id
@@ -235,14 +221,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location ### test.blobstore.path ###/table1
name default.table1
- numFiles 1
- numRows 1
- rawDataSize 1
- serialization.ddl struct table1 { i32 id}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
@@ -271,8 +251,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns id
@@ -281,14 +259,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location ### test.blobstore.path ###/table1
name default.table1
- numFiles 1
- numRows 1
- rawDataSize 1
- serialization.ddl struct table1 { i32 id}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
TotalFiles: 1
@@ -303,8 +275,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns id
@@ -313,21 +283,13 @@ STAGE PLANS:
#### A masked pattern was here ####
location ### test.blobstore.path ###/table1
name default.table1
- numFiles 1
- numRows 1
- rawDataSize 1
- serialization.ddl struct table1 { i32 id}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns id
@@ -336,14 +298,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location ### test.blobstore.path ###/table1
name default.table1
- numFiles 1
- numRows 1
- rawDataSize 1
- serialization.ddl struct table1 { i32 id}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
name: default.table1
@@ -359,8 +315,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns id
@@ -369,14 +323,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location ### test.blobstore.path ###/table1
name default.table1
- numFiles 1
- numRows 1
- rawDataSize 1
- serialization.ddl struct table1 { i32 id}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
@@ -395,8 +343,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns id
@@ -405,14 +351,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location ### test.blobstore.path ###/table1
name default.table1
- numFiles 1
- numRows 1
- rawDataSize 1
- serialization.ddl struct table1 { i32 id}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
TotalFiles: 1
@@ -427,8 +367,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns id
@@ -437,21 +375,13 @@ STAGE PLANS:
#### A masked pattern was here ####
location ### test.blobstore.path ###/table1
name default.table1
- numFiles 1
- numRows 1
- rawDataSize 1
- serialization.ddl struct table1 { i32 id}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns id
@@ -460,14 +390,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location ### test.blobstore.path ###/table1
name default.table1
- numFiles 1
- numRows 1
- rawDataSize 1
- serialization.ddl struct table1 { i32 id}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
name: default.table1
diff --git a/itests/hive-blobstore/src/test/results/clientpositive/write_final_output_blobstore.q.out b/itests/hive-blobstore/src/test/results/clientpositive/write_final_output_blobstore.q.out
index 2dd98ef..8bfc667 100644
--- a/itests/hive-blobstore/src/test/results/clientpositive/write_final_output_blobstore.q.out
+++ b/itests/hive-blobstore/src/test/results/clientpositive/write_final_output_blobstore.q.out
@@ -80,30 +80,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key
- columns.comments
columns.types int
#### A masked pattern was here ####
name default.hdfs_table
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct hdfs_table { i32 key}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key
@@ -111,14 +101,8 @@ STAGE PLANS:
columns.types int
#### A masked pattern was here ####
name default.hdfs_table
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct hdfs_table { i32 key}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.hdfs_table
name: default.hdfs_table
@@ -211,8 +195,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key
@@ -221,14 +203,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location ### test.blobstore.path ###/write_final_output_blobstore
name default.blobstore_table
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct blobstore_table { i32 key}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.blobstore_table
TotalFiles: 1
@@ -281,8 +257,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key
@@ -291,14 +265,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location ### test.blobstore.path ###/write_final_output_blobstore
name default.blobstore_table
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct blobstore_table { i32 key}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.blobstore_table
@@ -368,30 +336,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key
- columns.comments
columns.types int
#### A masked pattern was here ####
name default.hdfs_table
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct hdfs_table { i32 key}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key
@@ -399,14 +357,8 @@ STAGE PLANS:
columns.types int
#### A masked pattern was here ####
name default.hdfs_table
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct hdfs_table { i32 key}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.hdfs_table
name: default.hdfs_table
@@ -499,8 +451,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key
@@ -509,14 +459,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location ### test.blobstore.path ###/write_final_output_blobstore
name default.blobstore_table
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct blobstore_table { i32 key}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.blobstore_table
TotalFiles: 1
@@ -569,8 +513,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key
@@ -579,14 +521,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location ### test.blobstore.path ###/write_final_output_blobstore
name default.blobstore_table
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct blobstore_table { i32 key}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.blobstore_table
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/plan/PartitionDesc.java b/ql/src/java/org/apache/hadoop/hive/ql/plan/PartitionDesc.java
index b226ab7..fda8f46 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/plan/PartitionDesc.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/plan/PartitionDesc.java
@@ -73,8 +73,6 @@ public class PartitionDesc implements Serializable, Cloneable {
public PartitionDesc() {
}
- private final static org.slf4j.Logger LOG = org.slf4j.LoggerFactory.getLogger(PartitionDesc.class);
-
public PartitionDesc(final TableDesc table, final LinkedHashMap<String, String> partSpec) {
this.tableDesc = table;
setPartSpec(partSpec);
@@ -92,10 +90,6 @@ public class PartitionDesc implements Serializable, Cloneable {
}
}
- public PartitionDesc(final Partition part) throws HiveException {
- this(part, getTableDesc(part.getTable()));
- }
-
/**
* @param part Partition
* @param tblDesc Table Descriptor
@@ -223,6 +217,7 @@ public class PartitionDesc implements Serializable, Cloneable {
}
public void setProperties(final Properties properties) {
+ properties.remove("columns.comments");
if (properties instanceof CopyOnFirstWriteProperties) {
this.properties = properties;
} else {
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/plan/TableDesc.java b/ql/src/java/org/apache/hadoop/hive/ql/plan/TableDesc.java
index 7993779..0435477 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/plan/TableDesc.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/plan/TableDesc.java
@@ -45,8 +45,6 @@ import java.util.Properties;
*/
public class TableDesc implements Serializable, Cloneable {
- private static final Logger LOG = LoggerFactory.getLogger(TableDesc.class);
-
private static final long serialVersionUID = 1L;
private Class<? extends InputFormat> inputFileFormatClass;
private Class<? extends OutputFormat> outputFileFormatClass;
diff --git a/ql/src/test/results/clientpositive/beeline/smb_mapjoin_11.q.out b/ql/src/test/results/clientpositive/beeline/smb_mapjoin_11.q.out
index 4089597..5d09134 100644
--- a/ql/src/test/results/clientpositive/beeline/smb_mapjoin_11.q.out
+++ b/ql/src/test/results/clientpositive/beeline/smb_mapjoin_11.q.out
@@ -111,25 +111,17 @@ STAGE PLANS:
partition values:
ds 1
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
bucket_count 16
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.test_table1_n1
- numFiles 16
- numRows 500
partition_columns ds
partition_columns.types string
- rawDataSize 5312
- serialization.ddl struct test_table1_n1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -147,10 +139,8 @@ STAGE PLANS:
name default.test_table1_n1
partition_columns ds
partition_columns.types string
- serialization.ddl struct test_table1_n1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test_table1_n1
name: default.test_table1_n1
@@ -186,10 +176,8 @@ STAGE PLANS:
name default.test_table3_n1
partition_columns ds
partition_columns.types string
- serialization.ddl struct test_table3_n1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test_table3_n1
TotalFiles: 16
@@ -248,10 +236,8 @@ STAGE PLANS:
name default.test_table3_n1
partition_columns ds
partition_columns.types string
- serialization.ddl struct test_table3_n1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test_table3_n1
@@ -1977,25 +1963,17 @@ STAGE PLANS:
partition values:
ds 1
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
bucket_count 16
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.test_table1_n1
- numFiles 16
- numRows 500
partition_columns ds
partition_columns.types string
- rawDataSize 5312
- serialization.ddl struct test_table1_n1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -2013,10 +1991,8 @@ STAGE PLANS:
name default.test_table1_n1
partition_columns ds
partition_columns.types string
- serialization.ddl struct test_table1_n1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test_table1_n1
name: default.test_table1_n1
@@ -2096,25 +2072,17 @@ STAGE PLANS:
partition values:
ds 1
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
bucket_count 16
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.test_table3_n1
- numFiles 16
- numRows 1028
partition_columns ds
partition_columns.types string
- rawDataSize 10968
- serialization.ddl struct test_table3_n1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 11996
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -2131,10 +2099,8 @@ STAGE PLANS:
name default.test_table3_n1
partition_columns ds
partition_columns.types string
- serialization.ddl struct test_table3_n1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test_table3_n1
name: default.test_table3_n1
diff --git a/ql/src/test/results/clientpositive/beeline/smb_mapjoin_12.q.out b/ql/src/test/results/clientpositive/beeline/smb_mapjoin_12.q.out
index d7bd71a..bbe6b8b 100644
--- a/ql/src/test/results/clientpositive/beeline/smb_mapjoin_12.q.out
+++ b/ql/src/test/results/clientpositive/beeline/smb_mapjoin_12.q.out
@@ -129,10 +129,8 @@ STAGE PLANS:
name default.test_table3_n4
partition_columns ds
partition_columns.types string
- serialization.ddl struct test_table3_n4 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test_table3_n4
TotalFiles: 1
@@ -149,25 +147,17 @@ STAGE PLANS:
partition values:
ds 1
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
bucket_count 16
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.test_table1_n6
- numFiles 16
- numRows 500
partition_columns ds
partition_columns.types string
- rawDataSize 5312
- serialization.ddl struct test_table1_n6 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -185,10 +175,8 @@ STAGE PLANS:
name default.test_table1_n6
partition_columns ds
partition_columns.types string
- serialization.ddl struct test_table1_n6 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test_table1_n6
name: default.test_table1_n6
@@ -218,10 +206,8 @@ STAGE PLANS:
name default.test_table3_n4
partition_columns ds
partition_columns.types string
- serialization.ddl struct test_table3_n4 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test_table3_n4
@@ -343,25 +329,17 @@ STAGE PLANS:
partition values:
ds 1
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
bucket_count 16
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.test_table3_n4
- numFiles 16
- numRows 3084
partition_columns ds
partition_columns.types string
- rawDataSize 32904
- serialization.ddl struct test_table3_n4 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 35988
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -379,10 +357,8 @@ STAGE PLANS:
name default.test_table3_n4
partition_columns ds
partition_columns.types string
- serialization.ddl struct test_table3_n4 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test_table3_n4
name: default.test_table3_n4
@@ -419,10 +395,8 @@ STAGE PLANS:
name default.test_table3_n4
partition_columns ds
partition_columns.types string
- serialization.ddl struct test_table3_n4 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test_table3_n4
TotalFiles: 16
@@ -482,10 +456,8 @@ STAGE PLANS:
name default.test_table3_n4
partition_columns ds
partition_columns.types string
- serialization.ddl struct test_table3_n4 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test_table3_n4
diff --git a/ql/src/test/results/clientpositive/beeline/smb_mapjoin_13.q.out b/ql/src/test/results/clientpositive/beeline/smb_mapjoin_13.q.out
index 70033d8..4184920 100644
--- a/ql/src/test/results/clientpositive/beeline/smb_mapjoin_13.q.out
+++ b/ql/src/test/results/clientpositive/beeline/smb_mapjoin_13.q.out
@@ -124,31 +124,22 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
SORTBUCKETCOLSPREFIX TRUE
bucket_count 16
bucket_field_name key
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.test_table1_n12
- numFiles 16
- numRows 500
- rawDataSize 5312
- serialization.ddl struct test_table1_n12 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
SORTBUCKETCOLSPREFIX TRUE
bucket_count 16
bucket_field_name key
@@ -159,14 +150,8 @@ STAGE PLANS:
columns.types int:string
#### A masked pattern was here ####
name default.test_table1_n12
- numFiles 16
- numRows 500
- rawDataSize 5312
- serialization.ddl struct test_table1_n12 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test_table1_n12
name: default.test_table1_n12
@@ -322,31 +307,22 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
SORTBUCKETCOLSPREFIX TRUE
bucket_count 16
bucket_field_name key,value
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.test_table3_n6
- numFiles 16
- numRows 500
- rawDataSize 5312
- serialization.ddl struct test_table3_n6 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
SORTBUCKETCOLSPREFIX TRUE
bucket_count 16
bucket_field_name key,value
@@ -357,14 +333,8 @@ STAGE PLANS:
columns.types int:string
#### A masked pattern was here ####
name default.test_table3_n6
- numFiles 16
- numRows 500
- rawDataSize 5312
- serialization.ddl struct test_table3_n6 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test_table3_n6
name: default.test_table3_n6
diff --git a/ql/src/test/results/clientpositive/binary_output_format.q.out b/ql/src/test/results/clientpositive/binary_output_format.q.out
index b414360..760d863 100644
--- a/ql/src/test/results/clientpositive/binary_output_format.q.out
+++ b/ql/src/test/results/clientpositive/binary_output_format.q.out
@@ -101,8 +101,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"mydata":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns mydata
@@ -110,15 +108,9 @@ STAGE PLANS:
columns.types string
#### A masked pattern was here ####
name default.dest1_n109
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct dest1_n109 { string mydata}
serialization.format 1
serialization.last.column.takes.rest true
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1_n109
TotalFiles: 1
@@ -152,30 +144,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.src
- numFiles 1
- numRows 500
- rawDataSize 5312
- serialization.ddl struct src { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -183,14 +165,8 @@ STAGE PLANS:
columns.types string:string
#### A masked pattern was here ####
name default.src
- numFiles 1
- numRows 500
- rawDataSize 5312
- serialization.ddl struct src { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.src
name: default.src
@@ -246,8 +222,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"mydata":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns mydata
@@ -255,15 +229,9 @@ STAGE PLANS:
columns.types string
#### A masked pattern was here ####
name default.dest1_n109
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct dest1_n109 { string mydata}
serialization.format 1
serialization.last.column.takes.rest true
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1_n109
@@ -292,8 +260,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"mydata":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns mydata
@@ -301,15 +267,9 @@ STAGE PLANS:
columns.types string
#### A masked pattern was here ####
name default.dest1_n109
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct dest1_n109 { string mydata}
serialization.format 1
serialization.last.column.takes.rest true
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1_n109
TotalFiles: 1
@@ -324,8 +284,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"mydata":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns mydata
@@ -333,22 +291,14 @@ STAGE PLANS:
columns.types string
#### A masked pattern was here ####
name default.dest1_n109
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct dest1_n109 { string mydata}
serialization.format 1
serialization.last.column.takes.rest true
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"mydata":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns mydata
@@ -356,15 +306,9 @@ STAGE PLANS:
columns.types string
#### A masked pattern was here ####
name default.dest1_n109
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct dest1_n109 { string mydata}
serialization.format 1
serialization.last.column.takes.rest true
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1_n109
name: default.dest1_n109
@@ -386,8 +330,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"mydata":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns mydata
@@ -395,15 +337,9 @@ STAGE PLANS:
columns.types string
#### A masked pattern was here ####
name default.dest1_n109
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct dest1_n109 { string mydata}
serialization.format 1
serialization.last.column.takes.rest true
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1_n109
TotalFiles: 1
@@ -418,8 +354,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"mydata":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns mydata
@@ -427,22 +361,14 @@ STAGE PLANS:
columns.types string
#### A masked pattern was here ####
name default.dest1_n109
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct dest1_n109 { string mydata}
serialization.format 1
serialization.last.column.takes.rest true
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"mydata":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns mydata
@@ -450,15 +376,9 @@ STAGE PLANS:
columns.types string
#### A masked pattern was here ####
name default.dest1_n109
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct dest1_n109 { string mydata}
serialization.format 1
serialization.last.column.takes.rest true
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1_n109
name: default.dest1_n109
diff --git a/ql/src/test/results/clientpositive/bucket_map_join_1.q.out b/ql/src/test/results/clientpositive/bucket_map_join_1.q.out
index 440345f..9d72125 100644
--- a/ql/src/test/results/clientpositive/bucket_map_join_1.q.out
+++ b/ql/src/test/results/clientpositive/bucket_map_join_1.q.out
@@ -136,18 +136,11 @@ STAGE PLANS:
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.table1_n9
- numFiles 1
- numRows 0
- rawDataSize 0
- serialization.ddl struct table1_n9 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 20
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -163,14 +156,8 @@ STAGE PLANS:
columns.types string:string
#### A masked pattern was here ####
name default.table1_n9
- numFiles 1
- numRows 0
- rawDataSize 0
- serialization.ddl struct table1_n9 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 20
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1_n9
name: default.table1_n9
diff --git a/ql/src/test/results/clientpositive/bucket_map_join_2.q.out b/ql/src/test/results/clientpositive/bucket_map_join_2.q.out
index 20b2703..6e7414c 100644
--- a/ql/src/test/results/clientpositive/bucket_map_join_2.q.out
+++ b/ql/src/test/results/clientpositive/bucket_map_join_2.q.out
@@ -136,18 +136,11 @@ STAGE PLANS:
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.table1
- numFiles 1
- numRows 0
- rawDataSize 0
- serialization.ddl struct table1 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 20
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -163,14 +156,8 @@ STAGE PLANS:
columns.types string:string
#### A masked pattern was here ####
name default.table1
- numFiles 1
- numRows 0
- rawDataSize 0
- serialization.ddl struct table1 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 20
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
name: default.table1
diff --git a/ql/src/test/results/clientpositive/bucket_map_join_spark1.q.out b/ql/src/test/results/clientpositive/bucket_map_join_spark1.q.out
index 98a45ee..83516bf 100644
--- a/ql/src/test/results/clientpositive/bucket_map_join_spark1.q.out
+++ b/ql/src/test/results/clientpositive/bucket_map_join_spark1.q.out
@@ -159,20 +159,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_n19
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_n19 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -189,10 +182,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_n19
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_n19 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_n19
name: default.srcbucket_mapjoin_part_n19
@@ -258,8 +249,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -267,14 +256,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n9
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result_n9 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n9
TotalFiles: 1
@@ -326,20 +309,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n16
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2_n16 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -356,10 +332,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n16
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n16 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n16
name: default.srcbucket_mapjoin_part_2_n16
@@ -375,20 +349,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_n19
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_n19 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -405,10 +372,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_n19
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_n19 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_n19
name: default.srcbucket_mapjoin_part_n19
@@ -424,8 +389,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -433,14 +396,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n9
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result_n9 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n9
@@ -627,20 +584,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_n19
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_n19 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -657,10 +607,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_n19
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_n19 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_n19
name: default.srcbucket_mapjoin_part_n19
@@ -726,8 +674,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -735,14 +681,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n9
- numFiles 1
- numRows 1028
- rawDataSize 19022
- serialization.ddl struct bucketmapjoin_tmp_result_n9 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 20050
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n9
TotalFiles: 1
@@ -794,20 +734,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n16
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2_n16 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -824,10 +757,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n16
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n16 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n16
name: default.srcbucket_mapjoin_part_2_n16
@@ -843,20 +774,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_n19
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_n19 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -873,10 +797,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_n19
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_n19 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_n19
name: default.srcbucket_mapjoin_part_n19
@@ -892,8 +814,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -901,14 +821,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n9
- numFiles 1
- numRows 1028
- rawDataSize 19022
- serialization.ddl struct bucketmapjoin_tmp_result_n9 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 20050
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n9
diff --git a/ql/src/test/results/clientpositive/bucket_map_join_spark2.q.out b/ql/src/test/results/clientpositive/bucket_map_join_spark2.q.out
index 902c129..c95804c 100644
--- a/ql/src/test/results/clientpositive/bucket_map_join_spark2.q.out
+++ b/ql/src/test/results/clientpositive/bucket_map_join_spark2.q.out
@@ -143,20 +143,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n10
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2_n10 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 3062
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -173,10 +166,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n10
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n10 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n10
name: default.srcbucket_mapjoin_part_2_n10
@@ -242,8 +233,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -251,14 +240,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n5
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result_n5 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n5
TotalFiles: 1
@@ -310,20 +293,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n10
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2_n10 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 3062
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -340,10 +316,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n10
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n10 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n10
name: default.srcbucket_mapjoin_part_2_n10
@@ -359,20 +333,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_n12
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_n12 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -389,10 +356,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_n12
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_n12 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_n12
name: default.srcbucket_mapjoin_part_n12
@@ -408,8 +373,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -417,14 +380,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n5
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result_n5 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n5
@@ -611,20 +568,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n10
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2_n10 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 3062
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -641,10 +591,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n10
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n10 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n10
name: default.srcbucket_mapjoin_part_2_n10
@@ -710,8 +658,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -719,14 +665,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n5
- numFiles 1
- numRows 564
- rawDataSize 10503
- serialization.ddl struct bucketmapjoin_tmp_result_n5 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 11067
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n5
TotalFiles: 1
@@ -778,20 +718,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n10
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2_n10 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 3062
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -808,10 +741,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n10
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n10 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n10
name: default.srcbucket_mapjoin_part_2_n10
@@ -827,20 +758,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_n12
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_n12 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -857,10 +781,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_n12
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_n12 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_n12
name: default.srcbucket_mapjoin_part_n12
@@ -876,8 +798,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -885,14 +805,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n5
- numFiles 1
- numRows 564
- rawDataSize 10503
- serialization.ddl struct bucketmapjoin_tmp_result_n5 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 11067
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n5
diff --git a/ql/src/test/results/clientpositive/bucket_map_join_spark3.q.out b/ql/src/test/results/clientpositive/bucket_map_join_spark3.q.out
index 42a6998..bbff481 100644
--- a/ql/src/test/results/clientpositive/bucket_map_join_spark3.q.out
+++ b/ql/src/test/results/clientpositive/bucket_map_join_spark3.q.out
@@ -143,20 +143,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_n4
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_n4 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 3062
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -173,10 +166,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_n4
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_n4 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_n4
name: default.srcbucket_mapjoin_part_n4
@@ -242,8 +233,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -251,14 +240,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n1
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result_n1 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n1
TotalFiles: 1
@@ -310,20 +293,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n3
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2_n3 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -340,10 +316,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n3
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n3 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n3
name: default.srcbucket_mapjoin_part_2_n3
@@ -359,20 +333,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_n4
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_n4 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 3062
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -389,10 +356,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_n4
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_n4 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_n4
name: default.srcbucket_mapjoin_part_n4
@@ -408,8 +373,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -417,14 +380,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n1
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result_n1 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n1
@@ -611,20 +568,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_n4
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_n4 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 3062
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -641,10 +591,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_n4
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_n4 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_n4
name: default.srcbucket_mapjoin_part_n4
@@ -710,8 +658,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -719,14 +665,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n1
- numFiles 1
- numRows 564
- rawDataSize 10503
- serialization.ddl struct bucketmapjoin_tmp_result_n1 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 11067
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n1
TotalFiles: 1
@@ -778,20 +718,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n3
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2_n3 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -808,10 +741,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n3
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n3 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n3
name: default.srcbucket_mapjoin_part_2_n3
@@ -827,20 +758,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_n4
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_n4 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 3062
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -857,10 +781,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_n4
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_n4 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_n4
name: default.srcbucket_mapjoin_part_n4
@@ -876,8 +798,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -885,14 +805,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n1
- numFiles 1
- numRows 564
- rawDataSize 10503
- serialization.ddl struct bucketmapjoin_tmp_result_n1 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 11067
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n1
diff --git a/ql/src/test/results/clientpositive/bucket_map_join_spark4.q.out b/ql/src/test/results/clientpositive/bucket_map_join_spark4.q.out
index 5e6a28b..1528a78 100644
--- a/ql/src/test/results/clientpositive/bucket_map_join_spark4.q.out
+++ b/ql/src/test/results/clientpositive/bucket_map_join_spark4.q.out
@@ -212,31 +212,22 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
SORTBUCKETCOLSPREFIX TRUE
bucket_count 2
bucket_field_name key
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.tbl1_n0
- numFiles 2
- numRows 10
- rawDataSize 70
- serialization.ddl struct tbl1_n0 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 80
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
SORTBUCKETCOLSPREFIX TRUE
bucket_count 2
bucket_field_name key
@@ -247,14 +238,8 @@ STAGE PLANS:
columns.types int:string
#### A masked pattern was here ####
name default.tbl1_n0
- numFiles 2
- numRows 10
- rawDataSize 70
- serialization.ddl struct tbl1_n0 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 80
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.tbl1_n0
name: default.tbl1_n0
@@ -264,31 +249,22 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
SORTBUCKETCOLSPREFIX TRUE
bucket_count 2
bucket_field_name key
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.tbl2_n0
- numFiles 2
- numRows 10
- rawDataSize 70
- serialization.ddl struct tbl2_n0 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 80
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
SORTBUCKETCOLSPREFIX TRUE
bucket_count 2
bucket_field_name key
@@ -299,14 +275,8 @@ STAGE PLANS:
columns.types int:string
#### A masked pattern was here ####
name default.tbl2_n0
- numFiles 2
- numRows 10
- rawDataSize 70
- serialization.ddl struct tbl2_n0 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 80
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.tbl2_n0
name: default.tbl2_n0
@@ -316,31 +286,22 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
SORTBUCKETCOLSPREFIX TRUE
bucket_count 2
bucket_field_name key
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.tbl3
- numFiles 2
- numRows 10
- rawDataSize 70
- serialization.ddl struct tbl3 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 80
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
SORTBUCKETCOLSPREFIX TRUE
bucket_count 2
bucket_field_name key
@@ -351,14 +312,8 @@ STAGE PLANS:
columns.types int:string
#### A masked pattern was here ####
name default.tbl3
- numFiles 2
- numRows 10
- rawDataSize 70
- serialization.ddl struct tbl3 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 80
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.tbl3
name: default.tbl3
@@ -597,31 +552,22 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
SORTBUCKETCOLSPREFIX TRUE
bucket_count 2
bucket_field_name key
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.tbl1_n0
- numFiles 2
- numRows 10
- rawDataSize 70
- serialization.ddl struct tbl1_n0 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 80
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
SORTBUCKETCOLSPREFIX TRUE
bucket_count 2
bucket_field_name key
@@ -632,14 +578,8 @@ STAGE PLANS:
columns.types int:string
#### A masked pattern was here ####
name default.tbl1_n0
- numFiles 2
- numRows 10
- rawDataSize 70
- serialization.ddl struct tbl1_n0 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 80
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.tbl1_n0
name: default.tbl1_n0
@@ -649,31 +589,22 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
SORTBUCKETCOLSPREFIX TRUE
bucket_count 2
bucket_field_name key
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.tbl2_n0
- numFiles 2
- numRows 10
- rawDataSize 70
- serialization.ddl struct tbl2_n0 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 80
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
SORTBUCKETCOLSPREFIX TRUE
bucket_count 2
bucket_field_name key
@@ -684,14 +615,8 @@ STAGE PLANS:
columns.types int:string
#### A masked pattern was here ####
name default.tbl2_n0
- numFiles 2
- numRows 10
- rawDataSize 70
- serialization.ddl struct tbl2_n0 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 80
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.tbl2_n0
name: default.tbl2_n0
@@ -701,31 +626,22 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
SORTBUCKETCOLSPREFIX TRUE
bucket_count 2
bucket_field_name key
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.tbl3
- numFiles 2
- numRows 10
- rawDataSize 70
- serialization.ddl struct tbl3 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 80
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
SORTBUCKETCOLSPREFIX TRUE
bucket_count 2
bucket_field_name key
@@ -736,14 +652,8 @@ STAGE PLANS:
columns.types int:string
#### A masked pattern was here ####
name default.tbl3
- numFiles 2
- numRows 10
- rawDataSize 70
- serialization.ddl struct tbl3 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 80
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.tbl3
name: default.tbl3
diff --git a/ql/src/test/results/clientpositive/bucketcontext_1.q.out b/ql/src/test/results/clientpositive/bucketcontext_1.q.out
index 35b6ae8..515ec2b 100644
--- a/ql/src/test/results/clientpositive/bucketcontext_1.q.out
+++ b/ql/src/test/results/clientpositive/bucketcontext_1.q.out
@@ -140,20 +140,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_small_n14
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_small_n14 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -171,10 +164,8 @@ STAGE PLANS:
name default.bucket_small_n14
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_small_n14 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_small_n14
name: default.bucket_small_n14
@@ -256,20 +247,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n14
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n14 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -287,10 +271,8 @@ STAGE PLANS:
name default.bucket_big_n14
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n14 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n14
name: default.bucket_big_n14
@@ -306,20 +288,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n14
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n14 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -337,10 +312,8 @@ STAGE PLANS:
name default.bucket_big_n14
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n14 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n14
name: default.bucket_big_n14
@@ -474,20 +447,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n14
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n14 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -505,10 +471,8 @@ STAGE PLANS:
name default.bucket_big_n14
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n14 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n14
name: default.bucket_big_n14
@@ -524,20 +488,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n14
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n14 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -555,10 +512,8 @@ STAGE PLANS:
name default.bucket_big_n14
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n14 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n14
name: default.bucket_big_n14
diff --git a/ql/src/test/results/clientpositive/bucketcontext_2.q.out b/ql/src/test/results/clientpositive/bucketcontext_2.q.out
index 442e93b..6c5c251 100644
--- a/ql/src/test/results/clientpositive/bucketcontext_2.q.out
+++ b/ql/src/test/results/clientpositive/bucketcontext_2.q.out
@@ -124,20 +124,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_small_n8
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_small_n8 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -155,10 +148,8 @@ STAGE PLANS:
name default.bucket_small_n8
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_small_n8 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_small_n8
name: default.bucket_small_n8
@@ -240,20 +231,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n8
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n8 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -271,10 +255,8 @@ STAGE PLANS:
name default.bucket_big_n8
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n8 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n8
name: default.bucket_big_n8
@@ -290,20 +272,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n8
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n8 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -321,10 +296,8 @@ STAGE PLANS:
name default.bucket_big_n8
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n8 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n8
name: default.bucket_big_n8
@@ -458,20 +431,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n8
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n8 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -489,10 +455,8 @@ STAGE PLANS:
name default.bucket_big_n8
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n8 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n8
name: default.bucket_big_n8
@@ -508,20 +472,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n8
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n8 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -539,10 +496,8 @@ STAGE PLANS:
name default.bucket_big_n8
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n8 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n8
name: default.bucket_big_n8
diff --git a/ql/src/test/results/clientpositive/bucketcontext_3.q.out b/ql/src/test/results/clientpositive/bucketcontext_3.q.out
index 8fa490d..6358b46 100644
--- a/ql/src/test/results/clientpositive/bucketcontext_3.q.out
+++ b/ql/src/test/results/clientpositive/bucketcontext_3.q.out
@@ -124,20 +124,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_small_n4
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_small_n4 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -155,10 +148,8 @@ STAGE PLANS:
name default.bucket_small_n4
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_small_n4 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_small_n4
name: default.bucket_small_n4
@@ -172,20 +163,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_small_n4
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_small_n4 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -203,10 +187,8 @@ STAGE PLANS:
name default.bucket_small_n4
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_small_n4 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_small_n4
name: default.bucket_small_n4
@@ -288,20 +270,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n4
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n4 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -319,10 +294,8 @@ STAGE PLANS:
name default.bucket_big_n4
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n4 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n4
name: default.bucket_big_n4
@@ -455,20 +428,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n4
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n4 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -486,10 +452,8 @@ STAGE PLANS:
name default.bucket_big_n4
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n4 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n4
name: default.bucket_big_n4
diff --git a/ql/src/test/results/clientpositive/bucketcontext_4.q.out b/ql/src/test/results/clientpositive/bucketcontext_4.q.out
index c0f1017..74121e0 100644
--- a/ql/src/test/results/clientpositive/bucketcontext_4.q.out
+++ b/ql/src/test/results/clientpositive/bucketcontext_4.q.out
@@ -140,20 +140,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_small
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -171,10 +164,8 @@ STAGE PLANS:
name default.bucket_small
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_small
name: default.bucket_small
@@ -188,20 +179,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_small
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -219,10 +203,8 @@ STAGE PLANS:
name default.bucket_small
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_small
name: default.bucket_small
@@ -304,20 +286,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -335,10 +310,8 @@ STAGE PLANS:
name default.bucket_big
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big
name: default.bucket_big
@@ -471,20 +444,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -502,10 +468,8 @@ STAGE PLANS:
name default.bucket_big
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big
name: default.bucket_big
diff --git a/ql/src/test/results/clientpositive/bucketcontext_5.q.out b/ql/src/test/results/clientpositive/bucketcontext_5.q.out
index ac4496c..066adfd 100644
--- a/ql/src/test/results/clientpositive/bucketcontext_5.q.out
+++ b/ql/src/test/results/clientpositive/bucketcontext_5.q.out
@@ -166,18 +166,11 @@ STAGE PLANS:
bucketing_version 1
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n13
- numFiles 2
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucket_big_n13 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -193,14 +186,8 @@ STAGE PLANS:
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n13
- numFiles 2
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucket_big_n13 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n13
name: default.bucket_big_n13
@@ -321,18 +308,11 @@ STAGE PLANS:
bucketing_version 1
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n13
- numFiles 2
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucket_big_n13 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -348,14 +328,8 @@ STAGE PLANS:
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n13
- numFiles 2
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucket_big_n13 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n13
name: default.bucket_big_n13
diff --git a/ql/src/test/results/clientpositive/bucketcontext_6.q.out b/ql/src/test/results/clientpositive/bucketcontext_6.q.out
index 39ed5cc..5147d67 100644
--- a/ql/src/test/results/clientpositive/bucketcontext_6.q.out
+++ b/ql/src/test/results/clientpositive/bucketcontext_6.q.out
@@ -188,20 +188,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n7
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n7 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -219,10 +212,8 @@ STAGE PLANS:
name default.bucket_big_n7
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n7 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n7
name: default.bucket_big_n7
@@ -238,20 +229,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n7
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n7 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -269,10 +253,8 @@ STAGE PLANS:
name default.bucket_big_n7
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n7 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n7
name: default.bucket_big_n7
@@ -402,20 +384,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n7
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n7 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -433,10 +408,8 @@ STAGE PLANS:
name default.bucket_big_n7
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n7 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n7
name: default.bucket_big_n7
@@ -452,20 +425,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n7
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n7 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -483,10 +449,8 @@ STAGE PLANS:
name default.bucket_big_n7
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n7 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n7
name: default.bucket_big_n7
diff --git a/ql/src/test/results/clientpositive/bucketcontext_7.q.out b/ql/src/test/results/clientpositive/bucketcontext_7.q.out
index eb64514..f488e41 100644
--- a/ql/src/test/results/clientpositive/bucketcontext_7.q.out
+++ b/ql/src/test/results/clientpositive/bucketcontext_7.q.out
@@ -159,20 +159,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_small_n2
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_small_n2 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -190,10 +183,8 @@ STAGE PLANS:
name default.bucket_small_n2
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_small_n2 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_small_n2
name: default.bucket_small_n2
@@ -207,20 +198,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_small_n2
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_small_n2 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -238,10 +222,8 @@ STAGE PLANS:
name default.bucket_small_n2
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_small_n2 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_small_n2
name: default.bucket_small_n2
@@ -323,20 +305,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n2
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n2 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -354,10 +329,8 @@ STAGE PLANS:
name default.bucket_big_n2
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n2 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n2
name: default.bucket_big_n2
@@ -373,20 +346,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n2
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n2 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -404,10 +370,8 @@ STAGE PLANS:
name default.bucket_big_n2
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n2 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n2
name: default.bucket_big_n2
@@ -545,20 +509,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n2
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n2 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -576,10 +533,8 @@ STAGE PLANS:
name default.bucket_big_n2
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n2 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n2
name: default.bucket_big_n2
@@ -595,20 +550,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n2
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n2 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -626,10 +574,8 @@ STAGE PLANS:
name default.bucket_big_n2
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n2 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n2
name: default.bucket_big_n2
diff --git a/ql/src/test/results/clientpositive/bucketcontext_8.q.out b/ql/src/test/results/clientpositive/bucketcontext_8.q.out
index 245b961..1775947 100644
--- a/ql/src/test/results/clientpositive/bucketcontext_8.q.out
+++ b/ql/src/test/results/clientpositive/bucketcontext_8.q.out
@@ -159,20 +159,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_small_n10
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_small_n10 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -190,10 +183,8 @@ STAGE PLANS:
name default.bucket_small_n10
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_small_n10 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_small_n10
name: default.bucket_small_n10
@@ -207,20 +198,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_small_n10
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_small_n10 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -238,10 +222,8 @@ STAGE PLANS:
name default.bucket_small_n10
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_small_n10 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_small_n10
name: default.bucket_small_n10
@@ -323,20 +305,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n10
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n10 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -354,10 +329,8 @@ STAGE PLANS:
name default.bucket_big_n10
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n10 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n10
name: default.bucket_big_n10
@@ -373,20 +346,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n10
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n10 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -404,10 +370,8 @@ STAGE PLANS:
name default.bucket_big_n10
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n10 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n10
name: default.bucket_big_n10
@@ -545,20 +509,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n10
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n10 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -576,10 +533,8 @@ STAGE PLANS:
name default.bucket_big_n10
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n10 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n10
name: default.bucket_big_n10
@@ -595,20 +550,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n10
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n10 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -626,10 +574,8 @@ STAGE PLANS:
name default.bucket_big_n10
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n10 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n10
name: default.bucket_big_n10
diff --git a/ql/src/test/results/clientpositive/bucketmapjoin10.q.out b/ql/src/test/results/clientpositive/bucketmapjoin10.q.out
index c4d6070..c57f403 100644
--- a/ql/src/test/results/clientpositive/bucketmapjoin10.q.out
+++ b/ql/src/test/results/clientpositive/bucketmapjoin10.q.out
@@ -173,20 +173,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n13
- numFiles 3
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2_n13 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4200
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -203,10 +196,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n13
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n13 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n13
name: default.srcbucket_mapjoin_part_2_n13
@@ -220,20 +211,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n13
- numFiles 2
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2_n13 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -250,10 +234,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n13
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n13 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n13
name: default.srcbucket_mapjoin_part_2_n13
@@ -326,20 +308,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_1_n6
- numFiles 2
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_1_n6 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -356,10 +331,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_1_n6
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_1_n6 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_1_n6
name: default.srcbucket_mapjoin_part_1_n6
@@ -375,20 +348,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_1_n6
- numFiles 3
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_1_n6 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4200
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -405,10 +371,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_1_n6
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_1_n6 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_1_n6
name: default.srcbucket_mapjoin_part_1_n6
diff --git a/ql/src/test/results/clientpositive/bucketmapjoin11.q.out b/ql/src/test/results/clientpositive/bucketmapjoin11.q.out
index 16114c9..092b605 100644
--- a/ql/src/test/results/clientpositive/bucketmapjoin11.q.out
+++ b/ql/src/test/results/clientpositive/bucketmapjoin11.q.out
@@ -181,20 +181,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n6
- numFiles 4
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2_n6 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -211,10 +204,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n6
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n6 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n6
name: default.srcbucket_mapjoin_part_2_n6
@@ -228,20 +219,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n6
- numFiles 2
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2_n6 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -258,10 +242,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n6
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n6 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n6
name: default.srcbucket_mapjoin_part_2_n6
@@ -342,20 +324,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_1_n2
- numFiles 2
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_1_n2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -372,10 +347,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_1_n2
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_1_n2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_1_n2
name: default.srcbucket_mapjoin_part_1_n2
@@ -391,20 +364,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_1_n2
- numFiles 4
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_1_n2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -421,10 +387,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_1_n2
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_1_n2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_1_n2
name: default.srcbucket_mapjoin_part_1_n2
@@ -539,20 +503,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n6
- numFiles 4
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2_n6 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -569,10 +526,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n6
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n6 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n6
name: default.srcbucket_mapjoin_part_2_n6
@@ -586,20 +541,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n6
- numFiles 2
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2_n6 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -616,10 +564,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n6
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n6 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n6
name: default.srcbucket_mapjoin_part_2_n6
@@ -700,20 +646,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_1_n2
- numFiles 2
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_1_n2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -730,10 +669,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_1_n2
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_1_n2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_1_n2
name: default.srcbucket_mapjoin_part_1_n2
@@ -749,20 +686,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_1_n2
- numFiles 4
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_1_n2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -779,10 +709,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_1_n2
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_1_n2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_1_n2
name: default.srcbucket_mapjoin_part_1_n2
diff --git a/ql/src/test/results/clientpositive/bucketmapjoin12.q.out b/ql/src/test/results/clientpositive/bucketmapjoin12.q.out
index 5c45394..8a584b8 100644
--- a/ql/src/test/results/clientpositive/bucketmapjoin12.q.out
+++ b/ql/src/test/results/clientpositive/bucketmapjoin12.q.out
@@ -138,26 +138,18 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n0
- numFiles 2
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2_n0 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -167,10 +159,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n0
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n0 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n0
name: default.srcbucket_mapjoin_part_2_n0
@@ -251,20 +241,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_1
- numFiles 2
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -281,10 +264,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_1
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_1
name: default.srcbucket_mapjoin_part_1
@@ -386,23 +367,15 @@ STAGE PLANS:
partition values:
part 1
properties:
- bucket_count -1
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_3
- numFiles 2
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_3 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -419,10 +392,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_3
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_3 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_3
name: default.srcbucket_mapjoin_part_3
@@ -495,20 +466,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_1
- numFiles 2
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -525,10 +489,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_1
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_1
name: default.srcbucket_mapjoin_part_1
diff --git a/ql/src/test/results/clientpositive/bucketmapjoin13.q.out b/ql/src/test/results/clientpositive/bucketmapjoin13.q.out
index 693377f..77e986e 100644
--- a/ql/src/test/results/clientpositive/bucketmapjoin13.q.out
+++ b/ql/src/test/results/clientpositive/bucketmapjoin13.q.out
@@ -103,25 +103,17 @@ STAGE PLANS:
partition values:
part 1
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
bucket_count 2
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n18
- numFiles 2
- numRows 500
partition_columns part
partition_columns.types string
- rawDataSize 5312
- serialization.ddl struct srcbucket_mapjoin_part_2_n18 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -138,10 +130,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n18
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n18 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n18
name: default.srcbucket_mapjoin_part_2_n18
@@ -210,25 +200,17 @@ STAGE PLANS:
partition values:
part 1
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
bucket_count 2
bucket_field_name value
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_1_n8
- numFiles 2
- numRows 500
partition_columns part
partition_columns.types string
- rawDataSize 5312
- serialization.ddl struct srcbucket_mapjoin_part_1_n8 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -245,10 +227,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_1_n8
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_1_n8 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_1_n8
name: default.srcbucket_mapjoin_part_1_n8
@@ -260,25 +240,17 @@ STAGE PLANS:
partition values:
part 2
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
bucket_count 2
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_1_n8
- numFiles 2
- numRows 500
partition_columns part
partition_columns.types string
- rawDataSize 5312
- serialization.ddl struct srcbucket_mapjoin_part_1_n8 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -295,10 +267,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_1_n8
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_1_n8 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_1_n8
name: default.srcbucket_mapjoin_part_1_n8
@@ -403,25 +373,17 @@ STAGE PLANS:
partition values:
part 1
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
bucket_count 2
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n18
- numFiles 2
- numRows 500
partition_columns part
partition_columns.types string
- rawDataSize 5312
- serialization.ddl struct srcbucket_mapjoin_part_2_n18 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -438,10 +400,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n18
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n18 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n18
name: default.srcbucket_mapjoin_part_2_n18
@@ -518,25 +478,17 @@ STAGE PLANS:
partition values:
part 2
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
bucket_count 2
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_1_n8
- numFiles 2
- numRows 500
partition_columns part
partition_columns.types string
- rawDataSize 5312
- serialization.ddl struct srcbucket_mapjoin_part_1_n8 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -553,10 +505,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_1_n8
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_1_n8 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_1_n8
name: default.srcbucket_mapjoin_part_1_n8
@@ -666,25 +616,17 @@ STAGE PLANS:
partition values:
part 1
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
bucket_count 2
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n18
- numFiles 2
- numRows 500
partition_columns part
partition_columns.types string
- rawDataSize 5312
- serialization.ddl struct srcbucket_mapjoin_part_2_n18 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -701,10 +643,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n18
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n18 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n18
name: default.srcbucket_mapjoin_part_2_n18
@@ -781,25 +721,17 @@ STAGE PLANS:
partition values:
part 2
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
bucket_count 2
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_1_n8
- numFiles 2
- numRows 500
partition_columns part
partition_columns.types string
- rawDataSize 5312
- serialization.ddl struct srcbucket_mapjoin_part_1_n8 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -816,10 +748,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_1_n8
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_1_n8 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_1_n8
name: default.srcbucket_mapjoin_part_1_n8
@@ -929,25 +859,17 @@ STAGE PLANS:
partition values:
part 1
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
bucket_count 2
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n18
- numFiles 2
- numRows 500
partition_columns part
partition_columns.types string
- rawDataSize 5312
- serialization.ddl struct srcbucket_mapjoin_part_2_n18 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -964,10 +886,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n18
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n18 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n18
name: default.srcbucket_mapjoin_part_2_n18
@@ -1044,25 +964,17 @@ STAGE PLANS:
partition values:
part 2
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
bucket_count 2
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_1_n8
- numFiles 2
- numRows 500
partition_columns part
partition_columns.types string
- rawDataSize 5312
- serialization.ddl struct srcbucket_mapjoin_part_1_n8 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -1079,10 +991,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_1_n8
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_1_n8 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_1_n8
name: default.srcbucket_mapjoin_part_1_n8
diff --git a/ql/src/test/results/clientpositive/bucketmapjoin5.q.out b/ql/src/test/results/clientpositive/bucketmapjoin5.q.out
index b0e2931..b43c675 100644
--- a/ql/src/test/results/clientpositive/bucketmapjoin5.q.out
+++ b/ql/src/test/results/clientpositive/bucketmapjoin5.q.out
@@ -264,8 +264,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -273,14 +271,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result
TotalFiles: 1
@@ -322,20 +314,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_n0
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_n0 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -352,10 +337,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_n0
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_n0 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_n0
name: default.srcbucket_mapjoin_part_n0
@@ -371,20 +354,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_n0
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_n0 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -401,10 +377,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_n0
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_n0 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_n0
name: default.srcbucket_mapjoin_part_n0
@@ -461,8 +435,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -470,14 +442,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result
@@ -506,8 +472,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -515,14 +479,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result
TotalFiles: 1
@@ -537,8 +495,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -546,21 +502,13 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -568,14 +516,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result
name: default.bucketmapjoin_tmp_result
@@ -597,8 +539,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -606,14 +546,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result
TotalFiles: 1
@@ -628,8 +562,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -637,21 +569,13 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -659,14 +583,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result
name: default.bucketmapjoin_tmp_result
@@ -886,8 +804,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -895,14 +811,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
- numFiles 1
- numRows 928
- rawDataSize 17038
- serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 17966
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result
TotalFiles: 1
@@ -944,20 +854,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 3062
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -974,10 +877,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2
name: default.srcbucket_mapjoin_part_2
@@ -993,20 +894,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 3062
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -1023,10 +917,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2
name: default.srcbucket_mapjoin_part_2
@@ -1083,8 +975,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -1092,14 +982,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
- numFiles 1
- numRows 928
- rawDataSize 17038
- serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 17966
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result
@@ -1128,8 +1012,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -1137,14 +1019,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
- numFiles 1
- numRows 928
- rawDataSize 17038
- serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 17966
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result
TotalFiles: 1
@@ -1159,8 +1035,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -1168,21 +1042,13 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
- numFiles 1
- numRows 928
- rawDataSize 17038
- serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 17966
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -1190,14 +1056,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
- numFiles 1
- numRows 928
- rawDataSize 17038
- serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 17966
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result
name: default.bucketmapjoin_tmp_result
@@ -1219,8 +1079,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -1228,14 +1086,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
- numFiles 1
- numRows 928
- rawDataSize 17038
- serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 17966
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result
TotalFiles: 1
@@ -1250,8 +1102,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -1259,21 +1109,13 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
- numFiles 1
- numRows 928
- rawDataSize 17038
- serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 17966
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -1281,14 +1123,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
- numFiles 1
- numRows 928
- rawDataSize 17038
- serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 17966
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result
name: default.bucketmapjoin_tmp_result
diff --git a/ql/src/test/results/clientpositive/bucketmapjoin8.q.out b/ql/src/test/results/clientpositive/bucketmapjoin8.q.out
index e1b658e..e70f877 100644
--- a/ql/src/test/results/clientpositive/bucketmapjoin8.q.out
+++ b/ql/src/test/results/clientpositive/bucketmapjoin8.q.out
@@ -103,20 +103,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n4
- numFiles 2
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2_n4 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -133,10 +126,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n4
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n4 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n4
name: default.srcbucket_mapjoin_part_2_n4
@@ -217,20 +208,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_1_n1
- numFiles 2
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_1_n1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -247,10 +231,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_1_n1
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_1_n1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_1_n1
name: default.srcbucket_mapjoin_part_1_n1
@@ -364,20 +346,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n4
- numFiles 2
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2_n4 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -394,10 +369,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n4
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n4 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n4
name: default.srcbucket_mapjoin_part_2_n4
@@ -478,20 +451,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_1_n1
- numFiles 2
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_1_n1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -508,10 +474,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_1_n1
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_1_n1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_1_n1
name: default.srcbucket_mapjoin_part_1_n1
diff --git a/ql/src/test/results/clientpositive/bucketmapjoin9.q.out b/ql/src/test/results/clientpositive/bucketmapjoin9.q.out
index 709c780..49789b5 100644
--- a/ql/src/test/results/clientpositive/bucketmapjoin9.q.out
+++ b/ql/src/test/results/clientpositive/bucketmapjoin9.q.out
@@ -111,20 +111,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n12
- numFiles 3
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2_n12 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4200
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -141,10 +134,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n12
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n12 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n12
name: default.srcbucket_mapjoin_part_2_n12
@@ -217,20 +208,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_1_n5
- numFiles 2
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_1_n5 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -247,10 +231,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_1_n5
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_1_n5 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_1_n5
name: default.srcbucket_mapjoin_part_1_n5
@@ -397,20 +379,13 @@ STAGE PLANS:
bucket_field_name value
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n12
- numFiles 2
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2_n12 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -427,10 +402,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n12
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n12 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n12
name: default.srcbucket_mapjoin_part_2_n12
@@ -503,20 +476,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_1_n5
- numFiles 2
- numRows 0
partition_columns part
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_1_n5 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -533,10 +499,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_1_n5
partition_columns part
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_1_n5 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_1_n5
name: default.srcbucket_mapjoin_part_1_n5
diff --git a/ql/src/test/results/clientpositive/bucketmapjoin_negative.q.out b/ql/src/test/results/clientpositive/bucketmapjoin_negative.q.out
index c90fa59..bd2e6a8 100644
--- a/ql/src/test/results/clientpositive/bucketmapjoin_negative.q.out
+++ b/ql/src/test/results/clientpositive/bucketmapjoin_negative.q.out
@@ -112,20 +112,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_n10
- numFiles 3
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_n10 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4200
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -142,10 +135,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_n10
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_n10 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_n10
name: default.srcbucket_mapjoin_part_n10
@@ -203,8 +194,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -212,14 +201,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n4
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result_n4 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n4
TotalFiles: 1
@@ -260,18 +243,11 @@ STAGE PLANS:
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_n10
- numFiles 2
- numRows 0
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_n10 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -286,14 +262,8 @@ STAGE PLANS:
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_n10
- numFiles 2
- numRows 0
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_n10 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_n10
name: default.srcbucket_mapjoin_n10
@@ -349,8 +319,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -358,14 +326,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n4
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result_n4 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n4
@@ -394,8 +356,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -403,14 +363,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n4
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result_n4 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n4
TotalFiles: 1
@@ -425,8 +379,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -434,21 +386,13 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n4
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result_n4 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -456,14 +400,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n4
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result_n4 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n4
name: default.bucketmapjoin_tmp_result_n4
@@ -485,8 +423,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -494,14 +430,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n4
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result_n4 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n4
TotalFiles: 1
@@ -516,8 +446,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -525,21 +453,13 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n4
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result_n4 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -547,14 +467,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n4
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result_n4 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n4
name: default.bucketmapjoin_tmp_result_n4
diff --git a/ql/src/test/results/clientpositive/bucketmapjoin_negative2.q.out b/ql/src/test/results/clientpositive/bucketmapjoin_negative2.q.out
index 1adac25..cd3576a 100644
--- a/ql/src/test/results/clientpositive/bucketmapjoin_negative2.q.out
+++ b/ql/src/test/results/clientpositive/bucketmapjoin_negative2.q.out
@@ -123,20 +123,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n7
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2_n7 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 3062
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -153,10 +146,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n7
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n7 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n7
name: default.srcbucket_mapjoin_part_2_n7
@@ -170,20 +161,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part_2_n7
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_part_2_n7 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 3062
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -200,10 +184,8 @@ STAGE PLANS:
name default.srcbucket_mapjoin_part_2_n7
partition_columns ds
partition_columns.types string
- serialization.ddl struct srcbucket_mapjoin_part_2_n7 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_part_2_n7
name: default.srcbucket_mapjoin_part_2_n7
@@ -269,8 +251,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -278,14 +258,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n3
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result_n3 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n3
TotalFiles: 1
@@ -326,18 +300,11 @@ STAGE PLANS:
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_n5
- numFiles 2
- numRows 0
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_n5 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -352,14 +319,8 @@ STAGE PLANS:
columns.types int:string
#### A masked pattern was here ####
name default.srcbucket_mapjoin_n5
- numFiles 2
- numRows 0
- rawDataSize 0
- serialization.ddl struct srcbucket_mapjoin_n5 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2750
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket_mapjoin_n5
name: default.srcbucket_mapjoin_n5
@@ -415,8 +376,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -424,14 +383,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n3
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result_n3 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n3
@@ -460,8 +413,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -469,14 +420,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n3
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result_n3 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n3
TotalFiles: 1
@@ -491,8 +436,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -500,21 +443,13 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n3
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result_n3 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -522,14 +457,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n3
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result_n3 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n3
name: default.bucketmapjoin_tmp_result_n3
@@ -551,8 +480,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -560,14 +487,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n3
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result_n3 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n3
TotalFiles: 1
@@ -582,8 +503,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -591,21 +510,13 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n3
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result_n3 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value1,value2
@@ -613,14 +524,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result_n3
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct bucketmapjoin_tmp_result_n3 { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result_n3
name: default.bucketmapjoin_tmp_result_n3
diff --git a/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out b/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out
index 2958ae1..3e489ed 100644
--- a/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out
+++ b/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out
@@ -254,18 +254,11 @@ STAGE PLANS:
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.test1_n11
- numFiles 3
- numRows 0
- rawDataSize 0
- serialization.ddl struct test1_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4200
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -281,14 +274,8 @@ STAGE PLANS:
columns.types string:string
#### A masked pattern was here ####
name default.test1_n11
- numFiles 3
- numRows 0
- rawDataSize 0
- serialization.ddl struct test1_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4200
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test1_n11
name: default.test1_n11
@@ -413,18 +400,11 @@ STAGE PLANS:
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.test2_n7
- numFiles 3
- numRows 0
- rawDataSize 0
- serialization.ddl struct test2_n7 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4200
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -440,14 +420,8 @@ STAGE PLANS:
columns.types string:string
#### A masked pattern was here ####
name default.test2_n7
- numFiles 3
- numRows 0
- rawDataSize 0
- serialization.ddl struct test2_n7 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4200
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test2_n7
name: default.test2_n7
@@ -564,18 +538,11 @@ STAGE PLANS:
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.test1_n11
- numFiles 3
- numRows 0
- rawDataSize 0
- serialization.ddl struct test1_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4200
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -591,14 +558,8 @@ STAGE PLANS:
columns.types string:string
#### A masked pattern was here ####
name default.test1_n11
- numFiles 3
- numRows 0
- rawDataSize 0
- serialization.ddl struct test1_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4200
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test1_n11
name: default.test1_n11
@@ -717,18 +678,11 @@ STAGE PLANS:
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.test1_n11
- numFiles 3
- numRows 0
- rawDataSize 0
- serialization.ddl struct test1_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4200
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -744,14 +698,8 @@ STAGE PLANS:
columns.types string:string
#### A masked pattern was here ####
name default.test1_n11
- numFiles 3
- numRows 0
- rawDataSize 0
- serialization.ddl struct test1_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4200
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test1_n11
name: default.test1_n11
@@ -870,18 +818,11 @@ STAGE PLANS:
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.test1_n11
- numFiles 3
- numRows 0
- rawDataSize 0
- serialization.ddl struct test1_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4200
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -897,14 +838,8 @@ STAGE PLANS:
columns.types string:string
#### A masked pattern was here ####
name default.test1_n11
- numFiles 3
- numRows 0
- rawDataSize 0
- serialization.ddl struct test1_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4200
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test1_n11
name: default.test1_n11
@@ -1023,18 +958,11 @@ STAGE PLANS:
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.test1_n11
- numFiles 3
- numRows 0
- rawDataSize 0
- serialization.ddl struct test1_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4200
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -1050,14 +978,8 @@ STAGE PLANS:
columns.types string:string
#### A masked pattern was here ####
name default.test1_n11
- numFiles 3
- numRows 0
- rawDataSize 0
- serialization.ddl struct test1_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4200
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test1_n11
name: default.test1_n11
@@ -1176,18 +1098,11 @@ STAGE PLANS:
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.test2_n7
- numFiles 3
- numRows 0
- rawDataSize 0
- serialization.ddl struct test2_n7 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4200
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -1203,14 +1118,8 @@ STAGE PLANS:
columns.types string:string
#### A masked pattern was here ####
name default.test2_n7
- numFiles 3
- numRows 0
- rawDataSize 0
- serialization.ddl struct test2_n7 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4200
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test2_n7
name: default.test2_n7
@@ -1329,18 +1238,11 @@ STAGE PLANS:
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.test2_n7
- numFiles 3
- numRows 0
- rawDataSize 0
- serialization.ddl struct test2_n7 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4200
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -1356,14 +1258,8 @@ STAGE PLANS:
columns.types string:string
#### A masked pattern was here ####
name default.test2_n7
- numFiles 3
- numRows 0
- rawDataSize 0
- serialization.ddl struct test2_n7 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4200
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test2_n7
name: default.test2_n7
@@ -1482,18 +1378,11 @@ STAGE PLANS:
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.test3
- numFiles 3
- numRows 0
- rawDataSize 0
- serialization.ddl struct test3 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4200
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -1509,14 +1398,8 @@ STAGE PLANS:
columns.types string:string
#### A masked pattern was here ####
name default.test3
- numFiles 3
- numRows 0
- rawDataSize 0
- serialization.ddl struct test3 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4200
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test3
name: default.test3
diff --git a/ql/src/test/results/clientpositive/druid/druidkafkamini_basic.q.out b/ql/src/test/results/clientpositive/druid/druidkafkamini_basic.q.out
index 42e9e52..8d6ebe9 100644
--- a/ql/src/test/results/clientpositive/druid/druidkafkamini_basic.q.out
+++ b/ql/src/test/results/clientpositive/druid/druidkafkamini_basic.q.out
@@ -272,31 +272,21 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"fullname":"true","shortname":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns shortname,fullname
- columns.comments
columns.types string:string
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.languages
- numFiles 1
- numRows 2
- rawDataSize 20
- serialization.ddl struct languages { string shortname, string fullname}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 22
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"fullname":"true","shortname":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns shortname,fullname
@@ -305,14 +295,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.languages
- numFiles 1
- numRows 2
- rawDataSize 20
- serialization.ddl struct languages { string shortname, string fullname}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 22
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.languages
name: default.languages
@@ -351,13 +335,11 @@ STAGE PLANS:
input format: org.apache.hadoop.hive.druid.io.DruidQueryBasedInputFormat
output format: org.apache.hadoop.hive.druid.io.DruidOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"__time":"true","added":"true","deleted":"true","language":"true","page":"true","user":"true"}}
EXTERNAL TRUE
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns __time,page,user,language,added,deleted
- columns.comments
columns.types timestamp:string:string:string:int:int
druid.datasource default.druid_kafka_test
druid.fieldNames language,user
@@ -378,23 +360,15 @@ STAGE PLANS:
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.druid_kafka_test
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct druid_kafka_test { timestamp __time, string page, string user, string language, i32 added, i32 deleted}
serialization.format 1
serialization.lib org.apache.hadoop.hive.druid.serde.DruidSerDe
storage_handler org.apache.hadoop.hive.druid.DruidStorageHandler
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.druid.serde.DruidSerDe
input format: org.apache.hadoop.hive.druid.io.DruidQueryBasedInputFormat
output format: org.apache.hadoop.hive.druid.io.DruidOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"__time":"true","added":"true","deleted":"true","language":"true","page":"true","user":"true"}}
EXTERNAL TRUE
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns __time,page,user,language,added,deleted
@@ -419,15 +393,9 @@ STAGE PLANS:
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.druid_kafka_test
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct druid_kafka_test { timestamp __time, string page, string user, string language, i32 added, i32 deleted}
serialization.format 1
serialization.lib org.apache.hadoop.hive.druid.serde.DruidSerDe
storage_handler org.apache.hadoop.hive.druid.DruidStorageHandler
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.druid.serde.DruidSerDe
name: default.druid_kafka_test
name: default.druid_kafka_test
diff --git a/ql/src/test/results/clientpositive/druid/druidmini_expressions.q.out b/ql/src/test/results/clientpositive/druid/druidmini_expressions.q.out
index a8af291..67ba37f 100644
--- a/ql/src/test/results/clientpositive/druid/druidmini_expressions.q.out
+++ b/ql/src/test/results/clientpositive/druid/druidmini_expressions.q.out
@@ -240,13 +240,11 @@ STAGE PLANS:
input format: org.apache.hadoop.hive.druid.io.DruidQueryBasedInputFormat
output format: org.apache.hadoop.hive.druid.io.DruidOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
EXTERNAL TRUE
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns __time,cstring1,cstring2,cdouble,cfloat,ctinyint,csmallint,cint,cbigint,cboolean1,cboolean2,cintstring,cfloatstring,cdoublestring
- columns.comments
columns.types timestamp with local time zone:string:string:double:float:tinyint:smallint:int:bigint:boolean:boolean:string:string:string
druid.datasource default.druid_table_alltypesorc
druid.fieldNames vc
@@ -258,23 +256,15 @@ STAGE PLANS:
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.druid_table_alltypesorc
- numFiles 0
- numRows 9173
- rawDataSize 0
- serialization.ddl struct druid_table_alltypesorc { timestamp with local time zone __time, string cstring1, string cstring2, double cdouble, float cfloat, byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, bool cboolean1, bool cboolean2, string cintstring, string cfloatstring, string cdoublestring}
serialization.format 1
serialization.lib org.apache.hadoop.hive.druid.serde.DruidSerDe
storage_handler org.apache.hadoop.hive.druid.DruidStorageHandler
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.druid.serde.DruidSerDe
input format: org.apache.hadoop.hive.druid.io.DruidQueryBasedInputFormat
output format: org.apache.hadoop.hive.druid.io.DruidOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
EXTERNAL TRUE
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns __time,cstring1,cstring2,cdouble,cfloat,ctinyint,csmallint,cint,cbigint,cboolean1,cboolean2,cintstring,cfloatstring,cdoublestring
@@ -290,15 +280,9 @@ STAGE PLANS:
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.druid_table_alltypesorc
- numFiles 0
- numRows 9173
- rawDataSize 0
- serialization.ddl struct druid_table_alltypesorc { timestamp with local time zone __time, string cstring1, string cstring2, double cdouble, float cfloat, byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, bool cboolean1, bool cboolean2, string cintstring, string cfloatstring, string cdoublestring}
serialization.format 1
serialization.lib org.apache.hadoop.hive.druid.serde.DruidSerDe
storage_handler org.apache.hadoop.hive.druid.DruidStorageHandler
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.druid.serde.DruidSerDe
name: default.druid_table_alltypesorc
name: default.druid_table_alltypesorc
diff --git a/ql/src/test/results/clientpositive/encrypted/encryption_join_unencrypted_tbl.q.out b/ql/src/test/results/clientpositive/encrypted/encryption_join_unencrypted_tbl.q.out
index d6115d8..894750a 100644
--- a/ql/src/test/results/clientpositive/encrypted/encryption_join_unencrypted_tbl.q.out
+++ b/ql/src/test/results/clientpositive/encrypted/encryption_join_unencrypted_tbl.q.out
@@ -615,31 +615,21 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.encrypted_table
- numFiles 1
- numRows 500
- rawDataSize 5312
- serialization.ddl struct encrypted_table { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -648,14 +638,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.encrypted_table
- numFiles 1
- numRows 500
- rawDataSize 5312
- serialization.ddl struct encrypted_table { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.encrypted_table
name: default.encrypted_table
@@ -665,31 +649,21 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.src
- numFiles 1
- numRows 500
- rawDataSize 5312
- serialization.ddl struct src { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -698,14 +672,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.src
- numFiles 1
- numRows 500
- rawDataSize 5312
- serialization.ddl struct src { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.src
name: default.src
diff --git a/ql/src/test/results/clientpositive/encrypted/encryption_join_with_different_encryption_keys.q.out b/ql/src/test/results/clientpositive/encrypted/encryption_join_with_different_encryption_keys.q.out
index f6c7d50..9cd3b0d 100644
--- a/ql/src/test/results/clientpositive/encrypted/encryption_join_with_different_encryption_keys.q.out
+++ b/ql/src/test/results/clientpositive/encrypted/encryption_join_with_different_encryption_keys.q.out
@@ -133,31 +133,21 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.table_key_1
- numFiles 1
- numRows 500
- rawDataSize 5312
- serialization.ddl struct table_key_1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -166,14 +156,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.table_key_1
- numFiles 1
- numRows 500
- rawDataSize 5312
- serialization.ddl struct table_key_1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table_key_1
name: default.table_key_1
@@ -183,31 +167,21 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
- columns.comments
columns.types int:string
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.table_key_2
- numFiles 1
- numRows 500
- rawDataSize 5312
- serialization.ddl struct table_key_2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -216,14 +190,8 @@ STAGE PLANS:
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.table_key_2
- numFiles 1
- numRows 500
- rawDataSize 5312
- serialization.ddl struct table_key_2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table_key_2
name: default.table_key_2
diff --git a/ql/src/test/results/clientpositive/erasurecoding/erasure_explain.q.out b/ql/src/test/results/clientpositive/erasurecoding/erasure_explain.q.out
index adad89e..c948f8d 100644
--- a/ql/src/test/results/clientpositive/erasurecoding/erasure_explain.q.out
+++ b/ql/src/test/results/clientpositive/erasurecoding/erasure_explain.q.out
@@ -131,32 +131,22 @@ STAGE PLANS:
ds 2008-04-08
hr 11
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
- bucket_count -1
column.name.delimiter ,
columns key,value
- columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.srcpart
- numFiles 1
numFilesErasureCoded 1
- numRows 500
partition_columns ds/hr
partition_columns.types string:string
- rawDataSize 5312
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -167,10 +157,8 @@ STAGE PLANS:
name default.srcpart
partition_columns ds/hr
partition_columns.types string:string
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcpart
name: default.srcpart
@@ -181,32 +169,22 @@ STAGE PLANS:
ds 2008-04-08
hr 12
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
- bucket_count -1
column.name.delimiter ,
columns key,value
- columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.srcpart
- numFiles 1
numFilesErasureCoded 1
- numRows 500
partition_columns ds/hr
partition_columns.types string:string
- rawDataSize 5312
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -217,10 +195,8 @@ STAGE PLANS:
name default.srcpart
partition_columns ds/hr
partition_columns.types string:string
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcpart
name: default.srcpart
@@ -231,32 +207,22 @@ STAGE PLANS:
ds 2008-04-09
hr 11
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
- bucket_count -1
column.name.delimiter ,
columns key,value
- columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.srcpart
- numFiles 1
numFilesErasureCoded 1
- numRows 500
partition_columns ds/hr
partition_columns.types string:string
- rawDataSize 5312
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -267,10 +233,8 @@ STAGE PLANS:
name default.srcpart
partition_columns ds/hr
partition_columns.types string:string
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcpart
name: default.srcpart
@@ -281,32 +245,22 @@ STAGE PLANS:
ds 2008-04-09
hr 12
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
- bucket_count -1
column.name.delimiter ,
columns key,value
- columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.srcpart
- numFiles 1
numFilesErasureCoded 1
- numRows 500
partition_columns ds/hr
partition_columns.types string:string
- rawDataSize 5312
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -317,10 +271,8 @@ STAGE PLANS:
name default.srcpart
partition_columns ds/hr
partition_columns.types string:string
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcpart
name: default.srcpart
diff --git a/ql/src/test/results/clientpositive/join_map_ppr.q.out b/ql/src/test/results/clientpositive/join_map_ppr.q.out
index e3d8212..bf57f67 100644
--- a/ql/src/test/results/clientpositive/join_map_ppr.q.out
+++ b/ql/src/test/results/clientpositive/join_map_ppr.q.out
@@ -124,8 +124,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,val2
@@ -133,14 +131,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.dest_j1_n4
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct dest_j1_n4 { string key, string value, string val2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest_j1_n4
TotalFiles: 1
@@ -179,30 +171,20 @@ STAGE PLANS:
ds 2008-04-08
hr 11
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
- bucket_count -1
column.name.delimiter ,
columns key,value
- columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
- numFiles 1
- numRows 500
partition_columns ds/hr
partition_columns.types string:string
- rawDataSize 5312
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -212,10 +194,8 @@ STAGE PLANS:
name default.srcpart
partition_columns ds/hr
partition_columns.types string:string
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcpart
name: default.srcpart
@@ -271,8 +251,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,val2
@@ -280,14 +258,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.dest_j1_n4
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct dest_j1_n4 { string key, string value, string val2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest_j1_n4
@@ -316,8 +288,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,val2
@@ -325,14 +295,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.dest_j1_n4
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct dest_j1_n4 { string key, string value, string val2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest_j1_n4
TotalFiles: 1
@@ -347,8 +311,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,val2
@@ -356,21 +318,13 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.dest_j1_n4
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct dest_j1_n4 { string key, string value, string val2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,val2
@@ -378,14 +332,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.dest_j1_n4
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct dest_j1_n4 { string key, string value, string val2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest_j1_n4
name: default.dest_j1_n4
@@ -407,8 +355,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,val2
@@ -416,14 +362,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.dest_j1_n4
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct dest_j1_n4 { string key, string value, string val2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest_j1_n4
TotalFiles: 1
@@ -438,8 +378,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,val2
@@ -447,21 +385,13 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.dest_j1_n4
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct dest_j1_n4 { string key, string value, string val2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,val2
@@ -469,14 +399,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.dest_j1_n4
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct dest_j1_n4 { string key, string value, string val2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest_j1_n4
name: default.dest_j1_n4
@@ -783,8 +707,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,val2
@@ -792,14 +714,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.dest_j1_n4
- numFiles 1
- numRows 107
- rawDataSize 2018
- serialization.ddl struct dest_j1_n4 { string key, string value, string val2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2125
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest_j1_n4
TotalFiles: 1
@@ -838,30 +754,20 @@ STAGE PLANS:
ds 2008-04-08
hr 11
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
- bucket_count -1
column.name.delimiter ,
columns key,value
- columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
- numFiles 1
- numRows 500
partition_columns ds/hr
partition_columns.types string:string
- rawDataSize 5312
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -871,10 +777,8 @@ STAGE PLANS:
name default.srcpart
partition_columns ds/hr
partition_columns.types string:string
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcpart
name: default.srcpart
@@ -930,8 +834,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,val2
@@ -939,14 +841,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.dest_j1_n4
- numFiles 1
- numRows 107
- rawDataSize 2018
- serialization.ddl struct dest_j1_n4 { string key, string value, string val2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2125
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest_j1_n4
@@ -975,8 +871,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,val2
@@ -984,14 +878,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.dest_j1_n4
- numFiles 1
- numRows 107
- rawDataSize 2018
- serialization.ddl struct dest_j1_n4 { string key, string value, string val2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2125
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest_j1_n4
TotalFiles: 1
@@ -1006,8 +894,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,val2
@@ -1015,21 +901,13 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.dest_j1_n4
- numFiles 1
- numRows 107
- rawDataSize 2018
- serialization.ddl struct dest_j1_n4 { string key, string value, string val2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2125
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,val2
@@ -1037,14 +915,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.dest_j1_n4
- numFiles 1
- numRows 107
- rawDataSize 2018
- serialization.ddl struct dest_j1_n4 { string key, string value, string val2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2125
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest_j1_n4
name: default.dest_j1_n4
@@ -1066,8 +938,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,val2
@@ -1075,14 +945,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.dest_j1_n4
- numFiles 1
- numRows 107
- rawDataSize 2018
- serialization.ddl struct dest_j1_n4 { string key, string value, string val2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2125
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest_j1_n4
TotalFiles: 1
@@ -1097,8 +961,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,val2
@@ -1106,21 +968,13 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.dest_j1_n4
- numFiles 1
- numRows 107
- rawDataSize 2018
- serialization.ddl struct dest_j1_n4 { string key, string value, string val2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2125
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,val2
@@ -1128,14 +982,8 @@ STAGE PLANS:
columns.types string:string:string
#### A masked pattern was here ####
name default.dest_j1_n4
- numFiles 1
- numRows 107
- rawDataSize 2018
- serialization.ddl struct dest_j1_n4 { string key, string value, string val2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2125
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest_j1_n4
name: default.dest_j1_n4
diff --git a/ql/src/test/results/clientpositive/kafka/kafka_storage_handler.q.out b/ql/src/test/results/clientpositive/kafka/kafka_storage_handler.q.out
index 71af39c..75fb823 100644
--- a/ql/src/test/results/clientpositive/kafka/kafka_storage_handler.q.out
+++ b/ql/src/test/results/clientpositive/kafka/kafka_storage_handler.q.out
@@ -1195,7 +1195,6 @@ STAGE PLANS:
bucketing_version 2
column.name.delimiter ,
columns isrobot,channel,timestamp,flags,isunpatrolled,page,diffurl,added,comment,commentlength,isnew,isminor,delta,isanonymous,user,deltabucket,deleted,namespace,__key,__partition,__offset,__timestamp
- columns.comments 'from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer'
columns.types boolean:string:string:string:boolean:string:string:bigint:string:bigint:boolean:boolean:bigint:boolean:string:double:bigint:string:binary:int:bigint:bigint
#### A masked pattern was here ####
hive.kafka.max.retries 6
@@ -1209,15 +1208,9 @@ STAGE PLANS:
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.wiki_kafka_avro_table
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct wiki_kafka_avro_table { bool isrobot, string channel, string timestamp, string flags, bool isunpatrolled, string page, string diffurl, i64 added, string comment, i64 commentlength, bool isnew, bool isminor, i64 delta, bool isanonymous, string user, double deltabucket, i64 deleted, string namespace, binary __key, i32 __partition, i64 __offset, i64 __timestamp}
serialization.format 1
serialization.lib org.apache.hadoop.hive.kafka.KafkaSerDe
storage_handler org.apache.hadoop.hive.kafka.KafkaStorageHandler
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.kafka.KafkaSerDe
input format: org.apache.hadoop.hive.kafka.KafkaInputFormat
@@ -1294,7 +1287,6 @@ STAGE PLANS:
"type" : "string"
} ]
}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns isrobot,channel,timestamp,flags,isunpatrolled,page,diffurl,added,comment,commentlength,isnew,isminor,delta,isanonymous,user,deltabucket,deleted,namespace,__key,__partition,__offset,__timestamp
@@ -1312,15 +1304,9 @@ STAGE PLANS:
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.wiki_kafka_avro_table
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct wiki_kafka_avro_table { bool isrobot, string channel, string timestamp, string flags, bool isunpatrolled, string page, string diffurl, i64 added, string comment, i64 commentlength, bool isnew, bool isminor, i64 delta, bool isanonymous, string user, double deltabucket, i64 deleted, string namespace, binary __key, i32 __partition, i64 __offset, i64 __timestamp}
serialization.format 1
serialization.lib org.apache.hadoop.hive.kafka.KafkaSerDe
storage_handler org.apache.hadoop.hive.kafka.KafkaStorageHandler
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.kafka.KafkaSerDe
name: default.wiki_kafka_avro_table
name: default.wiki_kafka_avro_table
@@ -1522,7 +1508,6 @@ STAGE PLANS:
bucketing_version 2
column.name.delimiter ,
columns isrobot,channel,timestamp,flags,isunpatrolled,page,diffurl,added,comment,commentlength,isnew,isminor,delta,isanonymous,user,deltabucket,deleted,namespace,__key,__partition,__offset,__timestamp
- columns.comments 'from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer'
columns.types boolean:string:string:string:boolean:string:string:bigint:string:bigint:boolean:boolean:bigint:boolean:string:double:bigint:string:binary:int:bigint:bigint
#### A masked pattern was here ####
hive.kafka.max.retries 6
@@ -1536,15 +1521,9 @@ STAGE PLANS:
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.wiki_kafka_avro_table
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct wiki_kafka_avro_table { bool isrobot, string channel, string timestamp, string flags, bool isunpatrolled, string page, string diffurl, i64 added, string comment, i64 commentlength, bool isnew, bool isminor, i64 delta, bool isanonymous, string user, double deltabucket, i64 deleted, string namespace, binary __key, i32 __partition, i64 __offset, i64 __timestamp}
serialization.format 1
serialization.lib org.apache.hadoop.hive.kafka.KafkaSerDe
storage_handler org.apache.hadoop.hive.kafka.KafkaStorageHandler
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.kafka.KafkaSerDe
input format: org.apache.hadoop.hive.kafka.KafkaInputFormat
@@ -1621,7 +1600,6 @@ STAGE PLANS:
"type" : "string"
} ]
}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns isrobot,channel,timestamp,flags,isunpatrolled,page,diffurl,added,comment,commentlength,isnew,isminor,delta,isanonymous,user,deltabucket,deleted,namespace,__key,__partition,__offset,__timestamp
@@ -1639,15 +1617,9 @@ STAGE PLANS:
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.wiki_kafka_avro_table
- numFiles 0
- numRows 0
- rawDataSize 0
- serialization.ddl struct wiki_kafka_avro_table { bool isrobot, string channel, string timestamp, string flags, bool isunpatrolled, string page, string diffurl, i64 added, string comment, i64 commentlength, bool isnew, bool isminor, i64 delta, bool isanonymous, string user, double deltabucket, i64 deleted, string namespace, binary __key, i32 __partition, i64 __offset, i64 __timestamp}
serialization.format 1
serialization.lib org.apache.hadoop.hive.kafka.KafkaSerDe
storage_handler org.apache.hadoop.hive.kafka.KafkaStorageHandler
- totalSize 0
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.kafka.KafkaSerDe
name: default.wiki_kafka_avro_table
name: default.wiki_kafka_avro_table
diff --git a/ql/src/test/results/clientpositive/list_bucket_dml_8.q.out b/ql/src/test/results/clientpositive/list_bucket_dml_8.q.out
index 9851958..bc9bdc7 100644
--- a/ql/src/test/results/clientpositive/list_bucket_dml_8.q.out
+++ b/ql/src/test/results/clientpositive/list_bucket_dml_8.q.out
@@ -63,7 +63,6 @@ STAGE PLANS:
input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -73,10 +72,8 @@ STAGE PLANS:
name default.list_bucketing_dynamic_part_n2
partition_columns ds/hr
partition_columns.types string:string
- serialization.ddl struct list_bucketing_dynamic_part_n2 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
name: default.list_bucketing_dynamic_part_n2
TotalFiles: 1
@@ -116,30 +113,20 @@ STAGE PLANS:
ds 2008-04-08
hr 11
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
- bucket_count -1
column.name.delimiter ,
columns key,value
- columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
- numFiles 1
- numRows 500
partition_columns ds/hr
partition_columns.types string:string
- rawDataSize 5312
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -149,10 +136,8 @@ STAGE PLANS:
name default.srcpart
partition_columns ds/hr
partition_columns.types string:string
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcpart
name: default.srcpart
@@ -165,30 +150,20 @@ STAGE PLANS:
ds 2008-04-08
hr 12
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
- bucket_count -1
column.name.delimiter ,
columns key,value
- columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
- numFiles 1
- numRows 500
partition_columns ds/hr
partition_columns.types string:string
- rawDataSize 5312
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -198,10 +173,8 @@ STAGE PLANS:
name default.srcpart
partition_columns ds/hr
partition_columns.types string:string
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcpart
name: default.srcpart
@@ -257,7 +230,6 @@ STAGE PLANS:
input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -267,10 +239,8 @@ STAGE PLANS:
name default.list_bucketing_dynamic_part_n2
partition_columns ds/hr
partition_columns.types string:string
- serialization.ddl struct list_bucketing_dynamic_part_n2 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
name: default.list_bucketing_dynamic_part_n2
@@ -542,30 +512,20 @@ STAGE PLANS:
ds 2008-04-08
hr a1
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
- bucket_count -1
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.list_bucketing_dynamic_part_n2
- numFiles 2
- numRows 16
partition_columns ds/hr
partition_columns.types string:string
- rawDataSize 136
- serialization.ddl struct list_bucketing_dynamic_part_n2 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
- totalSize 310
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -575,10 +535,8 @@ STAGE PLANS:
name default.list_bucketing_dynamic_part_n2
partition_columns ds/hr
partition_columns.types string:string
- serialization.ddl struct list_bucketing_dynamic_part_n2 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
name: default.list_bucketing_dynamic_part_n2
name: default.list_bucketing_dynamic_part_n2
@@ -591,29 +549,20 @@ STAGE PLANS:
ds 2008-04-08
hr b1
properties:
- bucket_count -1
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.list_bucketing_dynamic_part_n2
- numFiles 3
- numRows 984
partition_columns ds/hr
partition_columns.types string:string
- rawDataSize 9488
- serialization.ddl struct list_bucketing_dynamic_part_n2 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
- totalSize 10586
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -623,10 +572,8 @@ STAGE PLANS:
name default.list_bucketing_dynamic_part_n2
partition_columns ds/hr
partition_columns.types string:string
- serialization.ddl struct list_bucketing_dynamic_part_n2 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
name: default.list_bucketing_dynamic_part_n2
name: default.list_bucketing_dynamic_part_n2
diff --git a/ql/src/test/results/clientpositive/llap/acid_bucket_pruning.q.out b/ql/src/test/results/clientpositive/llap/acid_bucket_pruning.q.out
index 81859fe..0e8f3af 100644
--- a/ql/src/test/results/clientpositive/llap/acid_bucket_pruning.q.out
+++ b/ql/src/test/results/clientpositive/llap/acid_bucket_pruning.q.out
@@ -97,33 +97,24 @@ STAGE PLANS:
input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true"}}
bucket_count 16
bucket_field_name a
bucketing_version 2
column.name.delimiter ,
columns a
- columns.comments
columns.types int
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.acidtbldefault
- numFiles 17
- numRows 9174
- rawDataSize 0
- serialization.ddl struct acidtbldefault { i32 a}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 34811
transactional true
transactional_properties default
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true"}}
bucket_count 16
bucket_field_name a
bucketing_version 2
@@ -134,16 +125,10 @@ STAGE PLANS:
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.acidtbldefault
- numFiles 17
- numRows 9174
- rawDataSize 0
- serialization.ddl struct acidtbldefault { i32 a}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 34811
transactional true
transactional_properties default
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.acidtbldefault
name: default.acidtbldefault
diff --git a/ql/src/test/results/clientpositive/llap/acid_nullscan.q.out b/ql/src/test/results/clientpositive/llap/acid_nullscan.q.out
index 85d58dd..b941555 100644
--- a/ql/src/test/results/clientpositive/llap/acid_nullscan.q.out
+++ b/ql/src/test/results/clientpositive/llap/acid_nullscan.q.out
@@ -89,32 +89,23 @@ STAGE PLANS:
input format: org.apache.hadoop.hive.ql.io.OneNullRowInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}}
bucket_count 2
bucket_field_name a
bucketing_version 2
column.name.delimiter ,
columns a,b
- columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.acid_vectorized_n1
- numFiles 3
- numRows 11
- rawDataSize 0
- serialization.ddl struct acid_vectorized_n1 { i32 a, string b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
- totalSize 2583
transactional true
transactional_properties default
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}}
bucket_count 2
bucket_field_name a
bucketing_version 2
@@ -124,16 +115,10 @@ STAGE PLANS:
columns.types int:string
#### A masked pattern was here ####
name default.acid_vectorized_n1
- numFiles 3
- numRows 11
- rawDataSize 0
- serialization.ddl struct acid_vectorized_n1 { i32 a, string b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2583
transactional true
transactional_properties default
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.acid_vectorized_n1
name: default.acid_vectorized_n1
diff --git a/ql/src/test/results/clientpositive/llap/alter_partition_coltype.q.out b/ql/src/test/results/clientpositive/llap/alter_partition_coltype.q.out
index afffbd2..3567e5c 100644
--- a/ql/src/test/results/clientpositive/llap/alter_partition_coltype.q.out
+++ b/ql/src/test/results/clientpositive/llap/alter_partition_coltype.q.out
@@ -263,30 +263,20 @@ STAGE PLANS:
dt 100
ts 3.0
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
- bucket_count -1
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.alter_coltype
- numFiles 1
- numRows 25
partition_columns dt/ts
partition_columns.types string:double
- rawDataSize 191
- serialization.ddl struct alter_coltype { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 216
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -296,10 +286,8 @@ STAGE PLANS:
name default.alter_coltype
partition_columns dt/ts
partition_columns.types string:double
- serialization.ddl struct alter_coltype { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.alter_coltype
name: default.alter_coltype
@@ -310,30 +298,20 @@ STAGE PLANS:
dt 100
ts 6.30
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
- bucket_count -1
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.alter_coltype
- numFiles 1
- numRows 25
partition_columns dt/ts
partition_columns.types string:double
- rawDataSize 191
- serialization.ddl struct alter_coltype { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 216
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -343,10 +321,8 @@ STAGE PLANS:
name default.alter_coltype
partition_columns dt/ts
partition_columns.types string:double
- serialization.ddl struct alter_coltype { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.alter_coltype
name: default.alter_coltype
@@ -507,30 +483,20 @@ STAGE PLANS:
partcol1 1
partcol2 1
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"intcol":"true"}}
- bucket_count -1
column.name.delimiter ,
columns intcol
- columns.comments
columns.types string
#### A masked pattern was here ####
name pt.alterdynamic_part_table
- numFiles 2
- numRows 2
partition_columns partcol1/partcol2
partition_columns.types int:string
- rawDataSize 3
- serialization.ddl struct alterdynamic_part_table { string intcol}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns intcol
@@ -540,10 +506,8 @@ STAGE PLANS:
name pt.alterdynamic_part_table
partition_columns partcol1/partcol2
partition_columns.types int:string
- serialization.ddl struct alterdynamic_part_table { string intcol}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: pt.alterdynamic_part_table
name: pt.alterdynamic_part_table
@@ -585,30 +549,20 @@ STAGE PLANS:
partcol1 2
partcol2 1
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"intcol":"true"}}
- bucket_count -1
column.name.delimiter ,
columns intcol
- columns.comments
columns.types string
#### A masked pattern was here ####
name pt.alterdynamic_part_table
- numFiles 1
- numRows 1
partition_columns partcol1/partcol2
partition_columns.types int:string
- rawDataSize 1
- serialization.ddl struct alterdynamic_part_table { string intcol}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 2
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns intcol
@@ -618,10 +572,8 @@ STAGE PLANS:
name pt.alterdynamic_part_table
partition_columns partcol1/partcol2
partition_columns.types int:string
- serialization.ddl struct alterdynamic_part_table { string intcol}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: pt.alterdynamic_part_table
name: pt.alterdynamic_part_table
diff --git a/ql/src/test/results/clientpositive/llap/analyze_table_null_partition.q.out b/ql/src/test/results/clientpositive/llap/analyze_table_null_partition.q.out
index 3206980..da06a7d 100644
--- a/ql/src/test/results/clientpositive/llap/analyze_table_null_partition.q.out
+++ b/ql/src/test/results/clientpositive/llap/analyze_table_null_partition.q.out
@@ -115,30 +115,20 @@ STAGE PLANS:
partition values:
age 15
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"name":"true"}}
- bucket_count -1
column.name.delimiter ,
columns name
- columns.comments
columns.types string
#### A masked pattern was here ####
name default.test2_n6
- numFiles 1
- numRows 1
partition_columns age
partition_columns.types int
- rawDataSize 3
- serialization.ddl struct test2_n6 { string name}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 4
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns name
@@ -148,10 +138,8 @@ STAGE PLANS:
name default.test2_n6
partition_columns age
partition_columns.types int
- serialization.ddl struct test2_n6 { string name}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test2_n6
name: default.test2_n6
@@ -161,30 +149,20 @@ STAGE PLANS:
partition values:
age 30
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"name":"true"}}
- bucket_count -1
column.name.delimiter ,
columns name
- columns.comments
columns.types string
#### A masked pattern was here ####
name default.test2_n6
- numFiles 1
- numRows 1
partition_columns age
partition_columns.types int
- rawDataSize 0
- serialization.ddl struct test2_n6 { string name}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 1
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns name
@@ -194,10 +172,8 @@ STAGE PLANS:
name default.test2_n6
partition_columns age
partition_columns.types int
- serialization.ddl struct test2_n6 { string name}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test2_n6
name: default.test2_n6
@@ -207,30 +183,20 @@ STAGE PLANS:
partition values:
age 40
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"name":"true"}}
- bucket_count -1
column.name.delimiter ,
columns name
- columns.comments
columns.types string
#### A masked pattern was here ####
name default.test2_n6
- numFiles 1
- numRows 1
partition_columns age
partition_columns.types int
- rawDataSize 4
- serialization.ddl struct test2_n6 { string name}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns name
@@ -240,10 +206,8 @@ STAGE PLANS:
name default.test2_n6
partition_columns age
partition_columns.types int
- serialization.ddl struct test2_n6 { string name}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test2_n6
name: default.test2_n6
@@ -253,30 +217,20 @@ STAGE PLANS:
partition values:
age __HIVE_DEFAULT_PARTITION__
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"name":"true"}}
- bucket_count -1
column.name.delimiter ,
columns name
- columns.comments
columns.types string
#### A masked pattern was here ####
name default.test2_n6
- numFiles 1
- numRows 2
partition_columns age
partition_columns.types int
- rawDataSize 4
- serialization.ddl struct test2_n6 { string name}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 6
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns name
@@ -286,10 +240,8 @@ STAGE PLANS:
name default.test2_n6
partition_columns age
partition_columns.types int
- serialization.ddl struct test2_n6 { string name}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.test2_n6
name: default.test2_n6
diff --git a/ql/src/test/results/clientpositive/llap/autoColumnStats_5a.q.out b/ql/src/test/results/clientpositive/llap/autoColumnStats_5a.q.out
index 24b40f0..e92048c 100644
--- a/ql/src/test/results/clientpositive/llap/autoColumnStats_5a.q.out
+++ b/ql/src/test/results/clientpositive/llap/autoColumnStats_5a.q.out
@@ -62,7 +62,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns a,b
@@ -72,10 +71,8 @@ STAGE PLANS:
name default.partitioned1
partition_columns part
partition_columns.types int
- serialization.ddl struct partitioned1 { i32 a, string b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.partitioned1
TotalFiles: 1
@@ -118,11 +115,9 @@ STAGE PLANS:
bucketing_version 2
column.name.delimiter ,
columns
- columns.comments
columns.types
#### A masked pattern was here ####
name _dummy_database._dummy_table
- serialization.ddl struct _dummy_table { }
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -130,7 +125,6 @@ STAGE PLANS:
input format: org.apache.hadoop.hive.ql.io.NullRowsInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns
@@ -138,7 +132,6 @@ STAGE PLANS:
columns.types
#### A masked pattern was here ####
name _dummy_database._dummy_table
- serialization.ddl struct _dummy_table { }
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -199,7 +192,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns a,b
@@ -209,10 +201,8 @@ STAGE PLANS:
name default.partitioned1
partition_columns part
partition_columns.types int
- serialization.ddl struct partitioned1 { i32 a, string b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.partitioned1
@@ -331,7 +321,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns a,b
@@ -341,10 +330,8 @@ STAGE PLANS:
name default.partitioned1
partition_columns part
partition_columns.types int
- serialization.ddl struct partitioned1 { i32 a, string b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.partitioned1
TotalFiles: 1
@@ -387,11 +374,9 @@ STAGE PLANS:
bucketing_version 2
column.name.delimiter ,
columns
- columns.comments
columns.types
#### A masked pattern was here ####
name _dummy_database._dummy_table
- serialization.ddl struct _dummy_table { }
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -399,7 +384,6 @@ STAGE PLANS:
input format: org.apache.hadoop.hive.ql.io.NullRowsInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns
@@ -407,7 +391,6 @@ STAGE PLANS:
columns.types
#### A masked pattern was here ####
name _dummy_database._dummy_table
- serialization.ddl struct _dummy_table { }
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -468,7 +451,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns a,b
@@ -478,10 +460,8 @@ STAGE PLANS:
name default.partitioned1
partition_columns part
partition_columns.types int
- serialization.ddl struct partitioned1 { i32 a, string b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.partitioned1
diff --git a/ql/src/test/results/clientpositive/llap/autoColumnStats_8.q.out b/ql/src/test/results/clientpositive/llap/autoColumnStats_8.q.out
index 3bb474c..48d1d92 100644
--- a/ql/src/test/results/clientpositive/llap/autoColumnStats_8.q.out
+++ b/ql/src/test/results/clientpositive/llap/autoColumnStats_8.q.out
@@ -98,7 +98,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -108,10 +107,8 @@ STAGE PLANS:
name default.nzhang_part8
partition_columns ds/hr
partition_columns.types string:string
- serialization.ddl struct nzhang_part8 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.nzhang_part8
TotalFiles: 1
@@ -160,7 +157,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -170,10 +166,8 @@ STAGE PLANS:
name default.nzhang_part8
partition_columns ds/hr
partition_columns.types string:string
- serialization.ddl struct nzhang_part8 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.nzhang_part8
TotalFiles: 1
@@ -215,30 +209,20 @@ STAGE PLANS:
ds 2008-04-08
hr 11
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
- bucket_count -1
column.name.delimiter ,
columns key,value
- columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
- numFiles 1
- numRows 500
partition_columns ds/hr
partition_columns.types string:string
- rawDataSize 5312
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -248,10 +232,8 @@ STAGE PLANS:
name default.srcpart
partition_columns ds/hr
partition_columns.types string:string
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcpart
name: default.srcpart
@@ -264,30 +246,20 @@ STAGE PLANS:
ds 2008-04-08
hr 12
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
- bucket_count -1
column.name.delimiter ,
columns key,value
- columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
- numFiles 1
- numRows 500
partition_columns ds/hr
partition_columns.types string:string
- rawDataSize 5312
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -297,10 +269,8 @@ STAGE PLANS:
name default.srcpart
partition_columns ds/hr
partition_columns.types string:string
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcpart
name: default.srcpart
@@ -313,30 +283,20 @@ STAGE PLANS:
ds 2008-04-09
hr 11
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
- bucket_count -1
column.name.delimiter ,
columns key,value
- columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
- numFiles 1
- numRows 500
partition_columns ds/hr
partition_columns.types string:string
- rawDataSize 5312
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -346,10 +306,8 @@ STAGE PLANS:
name default.srcpart
partition_columns ds/hr
partition_columns.types string:string
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcpart
name: default.srcpart
@@ -362,30 +320,20 @@ STAGE PLANS:
ds 2008-04-09
hr 12
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
- bucket_count -1
column.name.delimiter ,
columns key,value
- columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
- numFiles 1
- numRows 500
partition_columns ds/hr
partition_columns.types string:string
- rawDataSize 5312
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -395,10 +343,8 @@ STAGE PLANS:
name default.srcpart
partition_columns ds/hr
partition_columns.types string:string
- serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcpart
name: default.srcpart
@@ -499,7 +445,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -509,10 +454,8 @@ STAGE PLANS:
name default.nzhang_part8
partition_columns ds/hr
partition_columns.types string:string
- serialization.ddl struct nzhang_part8 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.nzhang_part8
@@ -533,7 +476,6 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
@@ -543,10 +485,8 @@ STAGE PLANS:
name default.nzhang_part8
partition_columns ds/hr
partition_columns.types string:string
- serialization.ddl struct nzhang_part8 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.nzhang_part8
diff --git a/ql/src/test/results/clientpositive/llap/auto_join_reordering_values.q.out b/ql/src/test/results/clientpositive/llap/auto_join_reordering_values.q.out
index e87dc54..18f394b 100644
--- a/ql/src/test/results/clientpositive/llap/auto_join_reordering_values.q.out
+++ b/ql/src/test/results/clientpositive/llap/auto_join_reordering_values.q.out
@@ -155,30 +155,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cityid":"true","date":"true","dealid":"true","time":"true","userid":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns dealid,date,time,cityid,userid
- columns.comments
columns.types int:string:string:int:int
#### A masked pattern was here ####
name default.orderpayment_small
- numFiles 1
- numRows 1
- rawDataSize 36
- serialization.ddl struct orderpayment_small { i32 dealid, string date, string time, i32 cityid, i32 userid}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 37
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cityid":"true","date":"true","dealid":"true","time":"true","userid":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns dealid,date,time,cityid,userid
@@ -186,14 +176,8 @@ STAGE PLANS:
columns.types int:string:string:int:int
#### A masked pattern was here ####
name default.orderpayment_small
- numFiles 1
- numRows 1
- rawDataSize 36
- serialization.ddl struct orderpayment_small { i32 dealid, string date, string time, i32 cityid, i32 userid}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 37
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.orderpayment_small
name: default.orderpayment_small
@@ -235,30 +219,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cityid":"true","date":"true","dealid":"true","time":"true","userid":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns dealid,date,time,cityid,userid
- columns.comments
columns.types int:string:string:int:int
#### A masked pattern was here ####
name default.orderpayment_small
- numFiles 1
- numRows 1
- rawDataSize 36
- serialization.ddl struct orderpayment_small { i32 dealid, string date, string time, i32 cityid, i32 userid}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 37
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cityid":"true","date":"true","dealid":"true","time":"true","userid":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns dealid,date,time,cityid,userid
@@ -266,14 +240,8 @@ STAGE PLANS:
columns.types int:string:string:int:int
#### A masked pattern was here ####
name default.orderpayment_small
- numFiles 1
- numRows 1
- rawDataSize 36
- serialization.ddl struct orderpayment_small { i32 dealid, string date, string time, i32 cityid, i32 userid}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 37
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.orderpayment_small
name: default.orderpayment_small
@@ -315,30 +283,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cityid":"true","date":"true","dealid":"true","time":"true","userid":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns dealid,date,time,cityid,userid
- columns.comments
columns.types int:string:string:int:int
#### A masked pattern was here ####
name default.orderpayment_small
- numFiles 1
- numRows 1
- rawDataSize 36
- serialization.ddl struct orderpayment_small { i32 dealid, string date, string time, i32 cityid, i32 userid}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 37
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cityid":"true","date":"true","dealid":"true","time":"true","userid":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns dealid,date,time,cityid,userid
@@ -346,14 +304,8 @@ STAGE PLANS:
columns.types int:string:string:int:int
#### A masked pattern was here ####
name default.orderpayment_small
- numFiles 1
- numRows 1
- rawDataSize 36
- serialization.ddl struct orderpayment_small { i32 dealid, string date, string time, i32 cityid, i32 userid}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 37
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.orderpayment_small
name: default.orderpayment_small
@@ -395,30 +347,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cityid":"true","date":"true","dealid":"true","time":"true","userid":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns dealid,date,time,cityid,userid
- columns.comments
columns.types int:string:string:int:int
#### A masked pattern was here ####
name default.orderpayment_small
- numFiles 1
- numRows 1
- rawDataSize 36
- serialization.ddl struct orderpayment_small { i32 dealid, string date, string time, i32 cityid, i32 userid}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 37
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cityid":"true","date":"true","dealid":"true","time":"true","userid":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns dealid,date,time,cityid,userid
@@ -426,14 +368,8 @@ STAGE PLANS:
columns.types int:string:string:int:int
#### A masked pattern was here ####
name default.orderpayment_small
- numFiles 1
- numRows 1
- rawDataSize 36
- serialization.ddl struct orderpayment_small { i32 dealid, string date, string time, i32 cityid, i32 userid}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 37
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.orderpayment_small
name: default.orderpayment_small
@@ -475,30 +411,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"userid":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns userid
- columns.comments
columns.types int
#### A masked pattern was here ####
name default.user_small
- numFiles 1
- numRows 100
- rawDataSize 288
- serialization.ddl struct user_small { i32 userid}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 388
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"userid":"true"}}
- bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns userid
@@ -506,14 +432,8 @@ STAGE PLANS:
columns.types int
#### A masked pattern was here ####
name default.user_small
- numFiles 1
- numRows 100
- rawDataSize 288
- serialization.ddl struct user_small { i32 userid}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 388
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.user_small
name: default.user_small
diff --git a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_1.q.out b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_1.q.out
index 096e225..7566b34 100644
--- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_1.q.out
+++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_1.q.out
@@ -168,20 +168,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_small_n1
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_small_n1 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 114
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -199,10 +192,8 @@ STAGE PLANS:
name default.bucket_small_n1
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_small_n1 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_small_n1
name: default.bucket_small_n1
@@ -260,20 +251,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n1
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n1 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -291,10 +275,8 @@ STAGE PLANS:
name default.bucket_big_n1
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n1 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n1
name: default.bucket_big_n1
@@ -309,20 +291,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n1
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n1 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -340,10 +315,8 @@ STAGE PLANS:
name default.bucket_big_n1
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n1 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n1
name: default.bucket_big_n1
@@ -472,20 +445,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_small_n1
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_small_n1 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 114
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -503,10 +469,8 @@ STAGE PLANS:
name default.bucket_small_n1
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_small_n1 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_small_n1
name: default.bucket_small_n1
@@ -564,20 +528,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n1
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n1 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -595,10 +552,8 @@ STAGE PLANS:
name default.bucket_big_n1
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n1 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n1
name: default.bucket_big_n1
@@ -613,20 +568,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n1
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n1 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -644,10 +592,8 @@ STAGE PLANS:
name default.bucket_big_n1
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n1 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n1
name: default.bucket_big_n1
@@ -776,20 +722,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_small_n1
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_small_n1 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 114
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -807,10 +746,8 @@ STAGE PLANS:
name default.bucket_small_n1
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_small_n1 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_small_n1
name: default.bucket_small_n1
@@ -868,20 +805,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n1
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n1 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -899,10 +829,8 @@ STAGE PLANS:
name default.bucket_big_n1
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n1 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n1
name: default.bucket_big_n1
@@ -917,20 +845,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n1
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n1 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -948,10 +869,8 @@ STAGE PLANS:
name default.bucket_big_n1
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n1 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n1
name: default.bucket_big_n1
diff --git a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_11.q.out b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_11.q.out
index 4d428a6..9902ff3 100644
--- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_11.q.out
+++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_11.q.out
@@ -179,20 +179,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_small_n11
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_small_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 114
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -210,10 +203,8 @@ STAGE PLANS:
name default.bucket_small_n11
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_small_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_small_n11
name: default.bucket_small_n11
@@ -261,20 +252,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n11
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -292,10 +276,8 @@ STAGE PLANS:
name default.bucket_big_n11
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n11
name: default.bucket_big_n11
@@ -311,20 +293,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n11
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -342,10 +317,8 @@ STAGE PLANS:
name default.bucket_big_n11
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n11
name: default.bucket_big_n11
@@ -501,20 +474,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_small_n11
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_small_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 114
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -532,10 +498,8 @@ STAGE PLANS:
name default.bucket_small_n11
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_small_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_small_n11
name: default.bucket_small_n11
@@ -593,20 +557,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n11
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -624,10 +581,8 @@ STAGE PLANS:
name default.bucket_big_n11
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n11
name: default.bucket_big_n11
@@ -642,20 +597,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n11
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -673,10 +621,8 @@ STAGE PLANS:
name default.bucket_big_n11
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n11
name: default.bucket_big_n11
@@ -794,20 +740,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_small_n11
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_small_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 114
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -825,10 +764,8 @@ STAGE PLANS:
name default.bucket_small_n11
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_small_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_small_n11
name: default.bucket_small_n11
@@ -882,20 +819,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n11
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -913,10 +843,8 @@ STAGE PLANS:
name default.bucket_big_n11
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n11
name: default.bucket_big_n11
@@ -931,20 +859,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n11
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -962,10 +883,8 @@ STAGE PLANS:
name default.bucket_big_n11
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n11
name: default.bucket_big_n11
@@ -1084,20 +1003,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_small_n11
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_small_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 114
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -1115,10 +1027,8 @@ STAGE PLANS:
name default.bucket_small_n11
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_small_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_small_n11
name: default.bucket_small_n11
@@ -1168,20 +1078,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n11
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -1199,10 +1102,8 @@ STAGE PLANS:
name default.bucket_big_n11
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n11
name: default.bucket_big_n11
@@ -1217,20 +1118,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n11
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -1248,10 +1142,8 @@ STAGE PLANS:
name default.bucket_big_n11
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n11
name: default.bucket_big_n11
@@ -1296,20 +1188,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n11
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -1327,10 +1212,8 @@ STAGE PLANS:
name default.bucket_big_n11
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n11
name: default.bucket_big_n11
@@ -1346,20 +1229,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n11
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -1377,10 +1253,8 @@ STAGE PLANS:
name default.bucket_big_n11
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n11 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n11
name: default.bucket_big_n11
diff --git a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_12.q.out b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_12.q.out
index cf73803..d0d01fe 100644
--- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_12.q.out
+++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_12.q.out
@@ -223,20 +223,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_medium
- numFiles 3
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_medium { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 170
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -254,10 +247,8 @@ STAGE PLANS:
name default.bucket_medium
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_medium { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_medium
name: default.bucket_medium
@@ -293,20 +284,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_medium
- numFiles 3
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_medium { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 170
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -324,10 +308,8 @@ STAGE PLANS:
name default.bucket_medium
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_medium { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_medium
name: default.bucket_medium
@@ -382,20 +364,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n15
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n15 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -413,10 +388,8 @@ STAGE PLANS:
name default.bucket_big_n15
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n15 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n15
name: default.bucket_big_n15
@@ -431,20 +404,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_big_n15
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_big_n15 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 5812
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -462,10 +428,8 @@ STAGE PLANS:
name default.bucket_big_n15
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_big_n15 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_big_n15
name: default.bucket_big_n15
@@ -514,20 +478,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_small_n15
- numFiles 2
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_small_n15 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 114
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -545,10 +502,8 @@ STAGE PLANS:
name default.bucket_small_n15
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_small_n15 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_small_n15
name: default.bucket_small_n15
diff --git a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out
index 60cfb52..fbdf2d6 100644
--- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out
+++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out
@@ -150,20 +150,13 @@ STAGE PLANS:
bucket_field_name key
column.name.delimiter ,
columns key,value
- columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.bucket_small_n3
- numFiles 4
- numRows 0
partition_columns ds
partition_columns.types string
- rawDataSize 0
- serialization.ddl struct bucket_small_n3 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- totalSize 226
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -181,10 +174,8 @@ STAGE PLANS:
name default.bucket_small_n3
partition_columns ds
partition_columns.types string
- serialization.ddl struct bucket_small_n3 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket_small_n3
name: default.bucket_small_n3
@@ -242,20 +233,13 @@ STAGE PLANS:
... 39819 lines suppressed ...