You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by pr...@apache.org on 2016/08/24 20:56:58 UTC

[04/51] [partial] hive git commit: HIVE-14553: Remove tez golden files after HIVE-14502 (Prasanth Jayachandran reviewed by Siddharth Seth)

http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/metadataonly1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/metadataonly1.q.out b/ql/src/test/results/clientpositive/tez/metadataonly1.q.out
deleted file mode 100644
index c79f34a..0000000
--- a/ql/src/test/results/clientpositive/tez/metadataonly1.q.out
+++ /dev/null
@@ -1,1948 +0,0 @@
-PREHOOK: query: CREATE TABLE TEST1(A INT, B DOUBLE) partitioned by (ds string)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@TEST1
-POSTHOOK: query: CREATE TABLE TEST1(A INT, B DOUBLE) partitioned by (ds string)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@TEST1
-PREHOOK: query: explain extended select max(ds) from TEST1
-PREHOOK: type: QUERY
-POSTHOOK: query: explain extended select max(ds) from TEST1
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Edges:
-        Reducer 2 <- Map 1 (SIMPLE_EDGE)
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: test1
-                  Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                  GatherStats: false
-                  Select Operator
-                    expressions: ds (type: string)
-                    outputColumnNames: ds
-                    Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                    Group By Operator
-                      aggregations: max(ds)
-                      mode: hash
-                      outputColumnNames: _col0
-                      Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                      Reduce Output Operator
-                        null sort order: 
-                        sort order: 
-                        Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                        tag: -1
-                        value expressions: _col0 (type: string)
-                        auto parallelism: false
-        Reducer 2 
-            Needs Tagging: false
-            Reduce Operator Tree:
-              Group By Operator
-                aggregations: max(VALUE._col0)
-                mode: mergepartial
-                outputColumnNames: _col0
-                Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-#### A masked pattern was here ####
-                  NumFilesPerFileSink: 1
-                  Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-#### A masked pattern was here ####
-                  table:
-                      input format: org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      properties:
-                        columns _col0
-                        columns.types string
-                        escape.delim \
-                        hive.serialization.extend.additional.nesting.levels true
-                        serialization.escape.crlf true
-                        serialization.format 1
-                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                  TotalFiles: 1
-                  GatherStats: false
-                  MultiFileSpray: false
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
-
-PREHOOK: query: select max(ds) from TEST1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@test1
-#### A masked pattern was here ####
-POSTHOOK: query: select max(ds) from TEST1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@test1
-#### A masked pattern was here ####
-NULL
-PREHOOK: query: alter table TEST1 add partition (ds='1')
-PREHOOK: type: ALTERTABLE_ADDPARTS
-PREHOOK: Output: default@test1
-POSTHOOK: query: alter table TEST1 add partition (ds='1')
-POSTHOOK: type: ALTERTABLE_ADDPARTS
-POSTHOOK: Output: default@test1
-POSTHOOK: Output: default@test1@ds=1
-PREHOOK: query: explain extended select max(ds) from TEST1
-PREHOOK: type: QUERY
-POSTHOOK: query: explain extended select max(ds) from TEST1
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Edges:
-        Reducer 2 <- Map 1 (SIMPLE_EDGE)
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: test1
-                  Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                  GatherStats: false
-                  Select Operator
-                    expressions: ds (type: string)
-                    outputColumnNames: ds
-                    Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                    Group By Operator
-                      aggregations: max(ds)
-                      mode: hash
-                      outputColumnNames: _col0
-                      Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                      Reduce Output Operator
-                        null sort order: 
-                        sort order: 
-                        Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                        tag: -1
-                        value expressions: _col0 (type: string)
-                        auto parallelism: false
-            Path -> Alias:
-              nullscan://null/default.test1/part_ds=1_ [test1]
-            Path -> Partition:
-              nullscan://null/default.test1/part_ds=1_ 
-                Partition
-                  input format: org.apache.hadoop.hive.ql.io.OneNullRowInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  partition values:
-                    ds 1
-                  properties:
-                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
-                    bucket_count -1
-                    columns a,b
-                    columns.comments 
-                    columns.types int:double
-#### A masked pattern was here ####
-                    name default.test1
-                    numFiles 0
-                    numRows 0
-                    partition_columns ds
-                    partition_columns.types string
-                    rawDataSize 0
-                    serialization.ddl struct test1 { i32 a, double b}
-                    serialization.format 1
-                    serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
-                    totalSize 0
-#### A masked pattern was here ####
-                  serde: org.apache.hadoop.hive.serde2.NullStructSerDe
-                
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns a,b
-                      columns.comments 
-                      columns.types int:double
-#### A masked pattern was here ####
-                      name default.test1
-                      partition_columns ds
-                      partition_columns.types string
-                      serialization.ddl struct test1 { i32 a, double b}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.test1
-                  name: default.test1
-            Truncated Path -> Alias:
-              nullscan://null/default.test1/part_ds=1_ [test1]
-        Reducer 2 
-            Needs Tagging: false
-            Reduce Operator Tree:
-              Group By Operator
-                aggregations: max(VALUE._col0)
-                mode: mergepartial
-                outputColumnNames: _col0
-                Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-#### A masked pattern was here ####
-                  NumFilesPerFileSink: 1
-                  Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-#### A masked pattern was here ####
-                  table:
-                      input format: org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      properties:
-                        columns _col0
-                        columns.types string
-                        escape.delim \
-                        hive.serialization.extend.additional.nesting.levels true
-                        serialization.escape.crlf true
-                        serialization.format 1
-                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                  TotalFiles: 1
-                  GatherStats: false
-                  MultiFileSpray: false
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
-
-PREHOOK: query: select max(ds) from TEST1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@test1
-PREHOOK: Input: default@test1@ds=1
-#### A masked pattern was here ####
-POSTHOOK: query: select max(ds) from TEST1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@test1
-POSTHOOK: Input: default@test1@ds=1
-#### A masked pattern was here ####
-1
-PREHOOK: query: explain extended select count(distinct ds) from TEST1
-PREHOOK: type: QUERY
-POSTHOOK: query: explain extended select count(distinct ds) from TEST1
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Edges:
-        Reducer 2 <- Map 1 (SIMPLE_EDGE)
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: test1
-                  Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                  GatherStats: false
-                  Select Operator
-                    expressions: ds (type: string)
-                    outputColumnNames: ds
-                    Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                    Group By Operator
-                      aggregations: count(DISTINCT ds)
-                      keys: ds (type: string)
-                      mode: hash
-                      outputColumnNames: _col0, _col1
-                      Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: COMPLETE
-                      Reduce Output Operator
-                        key expressions: _col0 (type: string)
-                        null sort order: a
-                        sort order: +
-                        Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: COMPLETE
-                        tag: -1
-                        auto parallelism: false
-            Path -> Alias:
-              nullscan://null/default.test1/part_ds=1_ [test1]
-            Path -> Partition:
-              nullscan://null/default.test1/part_ds=1_ 
-                Partition
-                  input format: org.apache.hadoop.hive.ql.io.OneNullRowInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  partition values:
-                    ds 1
-                  properties:
-                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
-                    bucket_count -1
-                    columns a,b
-                    columns.comments 
-                    columns.types int:double
-#### A masked pattern was here ####
-                    name default.test1
-                    numFiles 0
-                    numRows 0
-                    partition_columns ds
-                    partition_columns.types string
-                    rawDataSize 0
-                    serialization.ddl struct test1 { i32 a, double b}
-                    serialization.format 1
-                    serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
-                    totalSize 0
-#### A masked pattern was here ####
-                  serde: org.apache.hadoop.hive.serde2.NullStructSerDe
-                
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns a,b
-                      columns.comments 
-                      columns.types int:double
-#### A masked pattern was here ####
-                      name default.test1
-                      partition_columns ds
-                      partition_columns.types string
-                      serialization.ddl struct test1 { i32 a, double b}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.test1
-                  name: default.test1
-            Truncated Path -> Alias:
-              nullscan://null/default.test1/part_ds=1_ [test1]
-        Reducer 2 
-            Needs Tagging: false
-            Reduce Operator Tree:
-              Group By Operator
-                aggregations: count(DISTINCT KEY._col0:0._col0)
-                mode: mergepartial
-                outputColumnNames: _col0
-                Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-#### A masked pattern was here ####
-                  NumFilesPerFileSink: 1
-                  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
-#### A masked pattern was here ####
-                  table:
-                      input format: org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      properties:
-                        columns _col0
-                        columns.types bigint
-                        escape.delim \
-                        hive.serialization.extend.additional.nesting.levels true
-                        serialization.escape.crlf true
-                        serialization.format 1
-                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                  TotalFiles: 1
-                  GatherStats: false
-                  MultiFileSpray: false
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
-
-PREHOOK: query: select count(distinct ds) from TEST1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@test1
-PREHOOK: Input: default@test1@ds=1
-#### A masked pattern was here ####
-POSTHOOK: query: select count(distinct ds) from TEST1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@test1
-POSTHOOK: Input: default@test1@ds=1
-#### A masked pattern was here ####
-1
-PREHOOK: query: explain extended select count(ds) from TEST1
-PREHOOK: type: QUERY
-POSTHOOK: query: explain extended select count(ds) from TEST1
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Edges:
-        Reducer 2 <- Map 1 (SIMPLE_EDGE)
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: test1
-                  Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                  GatherStats: false
-                  Select Operator
-                    expressions: ds (type: string)
-                    outputColumnNames: ds
-                    Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                    Group By Operator
-                      aggregations: count(ds)
-                      mode: hash
-                      outputColumnNames: _col0
-                      Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
-                      Reduce Output Operator
-                        null sort order: 
-                        sort order: 
-                        Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
-                        tag: -1
-                        value expressions: _col0 (type: bigint)
-                        auto parallelism: false
-            Path -> Alias:
-#### A masked pattern was here ####
-            Path -> Partition:
-#### A masked pattern was here ####
-                Partition
-                  base file name: ds=1
-                  input format: org.apache.hadoop.mapred.TextInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  partition values:
-                    ds 1
-                  properties:
-                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
-                    bucket_count -1
-                    columns a,b
-                    columns.comments 
-                    columns.types int:double
-#### A masked pattern was here ####
-                    name default.test1
-                    numFiles 0
-                    numRows 0
-                    partition_columns ds
-                    partition_columns.types string
-                    rawDataSize 0
-                    serialization.ddl struct test1 { i32 a, double b}
-                    serialization.format 1
-                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    totalSize 0
-#### A masked pattern was here ####
-                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns a,b
-                      columns.comments 
-                      columns.types int:double
-#### A masked pattern was here ####
-                      name default.test1
-                      partition_columns ds
-                      partition_columns.types string
-                      serialization.ddl struct test1 { i32 a, double b}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.test1
-                  name: default.test1
-            Truncated Path -> Alias:
-              /test1/ds=1 [test1]
-        Reducer 2 
-            Needs Tagging: false
-            Reduce Operator Tree:
-              Group By Operator
-                aggregations: count(VALUE._col0)
-                mode: mergepartial
-                outputColumnNames: _col0
-                Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-#### A masked pattern was here ####
-                  NumFilesPerFileSink: 1
-                  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
-#### A masked pattern was here ####
-                  table:
-                      input format: org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      properties:
-                        columns _col0
-                        columns.types bigint
-                        escape.delim \
-                        hive.serialization.extend.additional.nesting.levels true
-                        serialization.escape.crlf true
-                        serialization.format 1
-                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                  TotalFiles: 1
-                  GatherStats: false
-                  MultiFileSpray: false
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
-
-PREHOOK: query: select count(ds) from TEST1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@test1
-PREHOOK: Input: default@test1@ds=1
-#### A masked pattern was here ####
-POSTHOOK: query: select count(ds) from TEST1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@test1
-POSTHOOK: Input: default@test1@ds=1
-#### A masked pattern was here ####
-0
-PREHOOK: query: alter table TEST1 add partition (ds='2')
-PREHOOK: type: ALTERTABLE_ADDPARTS
-PREHOOK: Output: default@test1
-POSTHOOK: query: alter table TEST1 add partition (ds='2')
-POSTHOOK: type: ALTERTABLE_ADDPARTS
-POSTHOOK: Output: default@test1
-POSTHOOK: Output: default@test1@ds=2
-PREHOOK: query: explain extended 
-select count(*) from TEST1 a2 join (select max(ds) m from TEST1) b on a2.ds=b.m
-PREHOOK: type: QUERY
-POSTHOOK: query: explain extended 
-select count(*) from TEST1 a2 join (select max(ds) m from TEST1) b on a2.ds=b.m
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Edges:
-        Reducer 2 <- Map 1 (SIMPLE_EDGE), Reducer 5 (SIMPLE_EDGE)
-        Reducer 3 <- Reducer 2 (SIMPLE_EDGE)
-        Reducer 5 <- Map 4 (SIMPLE_EDGE)
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: a2
-                  Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                  GatherStats: false
-                  Select Operator
-                    expressions: ds (type: string)
-                    outputColumnNames: _col0
-                    Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                    Reduce Output Operator
-                      key expressions: _col0 (type: string)
-                      null sort order: a
-                      sort order: +
-                      Map-reduce partition columns: _col0 (type: string)
-                      Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                      tag: 0
-                      auto parallelism: true
-            Path -> Alias:
-#### A masked pattern was here ####
-            Path -> Partition:
-#### A masked pattern was here ####
-                Partition
-                  base file name: ds=1
-                  input format: org.apache.hadoop.mapred.TextInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  partition values:
-                    ds 1
-                  properties:
-                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
-                    bucket_count -1
-                    columns a,b
-                    columns.comments 
-                    columns.types int:double
-#### A masked pattern was here ####
-                    name default.test1
-                    numFiles 0
-                    numRows 0
-                    partition_columns ds
-                    partition_columns.types string
-                    rawDataSize 0
-                    serialization.ddl struct test1 { i32 a, double b}
-                    serialization.format 1
-                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    totalSize 0
-#### A masked pattern was here ####
-                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns a,b
-                      columns.comments 
-                      columns.types int:double
-#### A masked pattern was here ####
-                      name default.test1
-                      partition_columns ds
-                      partition_columns.types string
-                      serialization.ddl struct test1 { i32 a, double b}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.test1
-                  name: default.test1
-#### A masked pattern was here ####
-                Partition
-                  base file name: ds=2
-                  input format: org.apache.hadoop.mapred.TextInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  partition values:
-                    ds 2
-                  properties:
-                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
-                    bucket_count -1
-                    columns a,b
-                    columns.comments 
-                    columns.types int:double
-#### A masked pattern was here ####
-                    name default.test1
-                    numFiles 0
-                    numRows 0
-                    partition_columns ds
-                    partition_columns.types string
-                    rawDataSize 0
-                    serialization.ddl struct test1 { i32 a, double b}
-                    serialization.format 1
-                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    totalSize 0
-#### A masked pattern was here ####
-                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns a,b
-                      columns.comments 
-                      columns.types int:double
-#### A masked pattern was here ####
-                      name default.test1
-                      partition_columns ds
-                      partition_columns.types string
-                      serialization.ddl struct test1 { i32 a, double b}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.test1
-                  name: default.test1
-            Truncated Path -> Alias:
-              /test1/ds=1 [a2]
-              /test1/ds=2 [a2]
-        Map 4 
-            Map Operator Tree:
-                TableScan
-                  alias: test1
-                  Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                  GatherStats: false
-                  Select Operator
-                    expressions: ds (type: string)
-                    outputColumnNames: ds
-                    Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                    Group By Operator
-                      aggregations: max(ds)
-                      mode: hash
-                      outputColumnNames: _col0
-                      Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                      Reduce Output Operator
-                        null sort order: 
-                        sort order: 
-                        Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                        tag: -1
-                        value expressions: _col0 (type: string)
-                        auto parallelism: false
-            Path -> Alias:
-#### A masked pattern was here ####
-            Path -> Partition:
-#### A masked pattern was here ####
-                Partition
-                  base file name: ds=1
-                  input format: org.apache.hadoop.mapred.TextInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  partition values:
-                    ds 1
-                  properties:
-                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
-                    bucket_count -1
-                    columns a,b
-                    columns.comments 
-                    columns.types int:double
-#### A masked pattern was here ####
-                    name default.test1
-                    numFiles 0
-                    numRows 0
-                    partition_columns ds
-                    partition_columns.types string
-                    rawDataSize 0
-                    serialization.ddl struct test1 { i32 a, double b}
-                    serialization.format 1
-                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    totalSize 0
-#### A masked pattern was here ####
-                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns a,b
-                      columns.comments 
-                      columns.types int:double
-#### A masked pattern was here ####
-                      name default.test1
-                      partition_columns ds
-                      partition_columns.types string
-                      serialization.ddl struct test1 { i32 a, double b}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.test1
-                  name: default.test1
-#### A masked pattern was here ####
-                Partition
-                  base file name: ds=2
-                  input format: org.apache.hadoop.mapred.TextInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  partition values:
-                    ds 2
-                  properties:
-                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
-                    bucket_count -1
-                    columns a,b
-                    columns.comments 
-                    columns.types int:double
-#### A masked pattern was here ####
-                    name default.test1
-                    numFiles 0
-                    numRows 0
-                    partition_columns ds
-                    partition_columns.types string
-                    rawDataSize 0
-                    serialization.ddl struct test1 { i32 a, double b}
-                    serialization.format 1
-                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    totalSize 0
-#### A masked pattern was here ####
-                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns a,b
-                      columns.comments 
-                      columns.types int:double
-#### A masked pattern was here ####
-                      name default.test1
-                      partition_columns ds
-                      partition_columns.types string
-                      serialization.ddl struct test1 { i32 a, double b}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.test1
-                  name: default.test1
-            Truncated Path -> Alias:
-              /test1/ds=1 [test1]
-              /test1/ds=2 [test1]
-        Reducer 2 
-            Needs Tagging: false
-            Reduce Operator Tree:
-              Merge Join Operator
-                condition map:
-                     Inner Join 0 to 1
-                keys:
-                  0 _col0 (type: string)
-                  1 _col0 (type: string)
-                Position of Big Table: 1
-                Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
-                Group By Operator
-                  aggregations: count()
-                  mode: hash
-                  outputColumnNames: _col0
-                  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
-                  Reduce Output Operator
-                    null sort order: 
-                    sort order: 
-                    Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
-                    tag: -1
-                    value expressions: _col0 (type: bigint)
-                    auto parallelism: false
-        Reducer 3 
-            Needs Tagging: false
-            Reduce Operator Tree:
-              Group By Operator
-                aggregations: count(VALUE._col0)
-                mode: mergepartial
-                outputColumnNames: _col0
-                Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-#### A masked pattern was here ####
-                  NumFilesPerFileSink: 1
-                  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
-#### A masked pattern was here ####
-                  table:
-                      input format: org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      properties:
-                        columns _col0
-                        columns.types bigint
-                        escape.delim \
-                        hive.serialization.extend.additional.nesting.levels true
-                        serialization.escape.crlf true
-                        serialization.format 1
-                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                  TotalFiles: 1
-                  GatherStats: false
-                  MultiFileSpray: false
-        Reducer 5 
-            Needs Tagging: false
-            Reduce Operator Tree:
-              Group By Operator
-                aggregations: max(VALUE._col0)
-                mode: mergepartial
-                outputColumnNames: _col0
-                Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                Filter Operator
-                  isSamplingPred: false
-                  predicate: _col0 is not null (type: boolean)
-                  Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                  Reduce Output Operator
-                    key expressions: _col0 (type: string)
-                    null sort order: a
-                    sort order: +
-                    Map-reduce partition columns: _col0 (type: string)
-                    Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                    tag: 1
-                    auto parallelism: true
-                  Select Operator
-                    expressions: _col0 (type: string)
-                    outputColumnNames: _col0
-                    Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                    Group By Operator
-                      keys: _col0 (type: string)
-                      mode: hash
-                      outputColumnNames: _col0
-                      Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                      Dynamic Partitioning Event Operator
-                        Target column: ds (string)
-                        Target Input: a2
-                        Partition key expr: ds
-                        Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                        Target Vertex: Map 1
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
-
-PREHOOK: query: select count(*) from TEST1 a2 join (select max(ds) m from TEST1) b on a2.ds=b.m
-PREHOOK: type: QUERY
-PREHOOK: Input: default@test1
-PREHOOK: Input: default@test1@ds=1
-PREHOOK: Input: default@test1@ds=2
-#### A masked pattern was here ####
-POSTHOOK: query: select count(*) from TEST1 a2 join (select max(ds) m from TEST1) b on a2.ds=b.m
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@test1
-POSTHOOK: Input: default@test1@ds=1
-POSTHOOK: Input: default@test1@ds=2
-#### A masked pattern was here ####
-0
-PREHOOK: query: CREATE TABLE TEST2(A INT, B DOUBLE) partitioned by (ds string, hr string)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@TEST2
-POSTHOOK: query: CREATE TABLE TEST2(A INT, B DOUBLE) partitioned by (ds string, hr string)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@TEST2
-PREHOOK: query: alter table TEST2 add partition (ds='1', hr='1')
-PREHOOK: type: ALTERTABLE_ADDPARTS
-PREHOOK: Output: default@test2
-POSTHOOK: query: alter table TEST2 add partition (ds='1', hr='1')
-POSTHOOK: type: ALTERTABLE_ADDPARTS
-POSTHOOK: Output: default@test2
-POSTHOOK: Output: default@test2@ds=1/hr=1
-PREHOOK: query: alter table TEST2 add partition (ds='1', hr='2')
-PREHOOK: type: ALTERTABLE_ADDPARTS
-PREHOOK: Output: default@test2
-POSTHOOK: query: alter table TEST2 add partition (ds='1', hr='2')
-POSTHOOK: type: ALTERTABLE_ADDPARTS
-POSTHOOK: Output: default@test2
-POSTHOOK: Output: default@test2@ds=1/hr=2
-PREHOOK: query: alter table TEST2 add partition (ds='1', hr='3')
-PREHOOK: type: ALTERTABLE_ADDPARTS
-PREHOOK: Output: default@test2
-POSTHOOK: query: alter table TEST2 add partition (ds='1', hr='3')
-POSTHOOK: type: ALTERTABLE_ADDPARTS
-POSTHOOK: Output: default@test2
-POSTHOOK: Output: default@test2@ds=1/hr=3
-PREHOOK: query: explain extended select ds, count(distinct hr) from TEST2 group by ds
-PREHOOK: type: QUERY
-POSTHOOK: query: explain extended select ds, count(distinct hr) from TEST2 group by ds
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Edges:
-        Reducer 2 <- Map 1 (SIMPLE_EDGE)
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: test2
-                  Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE
-                  GatherStats: false
-                  Select Operator
-                    expressions: ds (type: string), hr (type: string)
-                    outputColumnNames: ds, hr
-                    Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE
-                    Group By Operator
-                      aggregations: count(DISTINCT hr)
-                      keys: ds (type: string), hr (type: string)
-                      mode: hash
-                      outputColumnNames: _col0, _col1, _col2
-                      Statistics: Num rows: 1 Data size: 376 Basic stats: COMPLETE Column stats: COMPLETE
-                      Reduce Output Operator
-                        key expressions: _col0 (type: string), _col1 (type: string)
-                        null sort order: aa
-                        sort order: ++
-                        Map-reduce partition columns: _col0 (type: string)
-                        Statistics: Num rows: 1 Data size: 376 Basic stats: COMPLETE Column stats: COMPLETE
-                        tag: -1
-                        auto parallelism: true
-            Path -> Alias:
-              nullscan://null/default.test2/part_ds=1_hr=1_ [test2]
-              nullscan://null/default.test2/part_ds=1_hr=2_ [test2]
-              nullscan://null/default.test2/part_ds=1_hr=3_ [test2]
-            Path -> Partition:
-              nullscan://null/default.test2/part_ds=1_hr=1_ 
-                Partition
-                  input format: org.apache.hadoop.hive.ql.io.OneNullRowInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  partition values:
-                    ds 1
-                    hr 1
-                  properties:
-                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
-                    bucket_count -1
-                    columns a,b
-                    columns.comments 
-                    columns.types int:double
-#### A masked pattern was here ####
-                    name default.test2
-                    numFiles 0
-                    numRows 0
-                    partition_columns ds/hr
-                    partition_columns.types string:string
-                    rawDataSize 0
-                    serialization.ddl struct test2 { i32 a, double b}
-                    serialization.format 1
-                    serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
-                    totalSize 0
-#### A masked pattern was here ####
-                  serde: org.apache.hadoop.hive.serde2.NullStructSerDe
-                
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns a,b
-                      columns.comments 
-                      columns.types int:double
-#### A masked pattern was here ####
-                      name default.test2
-                      partition_columns ds/hr
-                      partition_columns.types string:string
-                      serialization.ddl struct test2 { i32 a, double b}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.test2
-                  name: default.test2
-              nullscan://null/default.test2/part_ds=1_hr=2_ 
-                Partition
-                  input format: org.apache.hadoop.hive.ql.io.OneNullRowInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  partition values:
-                    ds 1
-                    hr 2
-                  properties:
-                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
-                    bucket_count -1
-                    columns a,b
-                    columns.comments 
-                    columns.types int:double
-#### A masked pattern was here ####
-                    name default.test2
-                    numFiles 0
-                    numRows 0
-                    partition_columns ds/hr
-                    partition_columns.types string:string
-                    rawDataSize 0
-                    serialization.ddl struct test2 { i32 a, double b}
-                    serialization.format 1
-                    serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
-                    totalSize 0
-#### A masked pattern was here ####
-                  serde: org.apache.hadoop.hive.serde2.NullStructSerDe
-                
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns a,b
-                      columns.comments 
-                      columns.types int:double
-#### A masked pattern was here ####
-                      name default.test2
-                      partition_columns ds/hr
-                      partition_columns.types string:string
-                      serialization.ddl struct test2 { i32 a, double b}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.test2
-                  name: default.test2
-              nullscan://null/default.test2/part_ds=1_hr=3_ 
-                Partition
-                  input format: org.apache.hadoop.hive.ql.io.OneNullRowInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  partition values:
-                    ds 1
-                    hr 3
-                  properties:
-                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
-                    bucket_count -1
-                    columns a,b
-                    columns.comments 
-                    columns.types int:double
-#### A masked pattern was here ####
-                    name default.test2
-                    numFiles 0
-                    numRows 0
-                    partition_columns ds/hr
-                    partition_columns.types string:string
-                    rawDataSize 0
-                    serialization.ddl struct test2 { i32 a, double b}
-                    serialization.format 1
-                    serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
-                    totalSize 0
-#### A masked pattern was here ####
-                  serde: org.apache.hadoop.hive.serde2.NullStructSerDe
-                
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns a,b
-                      columns.comments 
-                      columns.types int:double
-#### A masked pattern was here ####
-                      name default.test2
-                      partition_columns ds/hr
-                      partition_columns.types string:string
-                      serialization.ddl struct test2 { i32 a, double b}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.test2
-                  name: default.test2
-            Truncated Path -> Alias:
-              nullscan://null/default.test2/part_ds=1_hr=1_ [test2]
-              nullscan://null/default.test2/part_ds=1_hr=2_ [test2]
-              nullscan://null/default.test2/part_ds=1_hr=3_ [test2]
-        Reducer 2 
-            Needs Tagging: false
-            Reduce Operator Tree:
-              Group By Operator
-                aggregations: count(DISTINCT KEY._col1:0._col0)
-                keys: KEY._col0 (type: string)
-                mode: mergepartial
-                outputColumnNames: _col0, _col1
-                Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: COMPLETE
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-#### A masked pattern was here ####
-                  NumFilesPerFileSink: 1
-                  Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: COMPLETE
-#### A masked pattern was here ####
-                  table:
-                      input format: org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      properties:
-                        columns _col0,_col1
-                        columns.types string:bigint
-                        escape.delim \
-                        hive.serialization.extend.additional.nesting.levels true
-                        serialization.escape.crlf true
-                        serialization.format 1
-                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                  TotalFiles: 1
-                  GatherStats: false
-                  MultiFileSpray: false
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
-
-PREHOOK: query: select ds, count(distinct hr) from TEST2 group by ds
-PREHOOK: type: QUERY
-PREHOOK: Input: default@test2
-PREHOOK: Input: default@test2@ds=1/hr=1
-PREHOOK: Input: default@test2@ds=1/hr=2
-PREHOOK: Input: default@test2@ds=1/hr=3
-#### A masked pattern was here ####
-POSTHOOK: query: select ds, count(distinct hr) from TEST2 group by ds
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@test2
-POSTHOOK: Input: default@test2@ds=1/hr=1
-POSTHOOK: Input: default@test2@ds=1/hr=2
-POSTHOOK: Input: default@test2@ds=1/hr=3
-#### A masked pattern was here ####
-1	3
-PREHOOK: query: explain extended select ds, count(hr) from TEST2 group by ds
-PREHOOK: type: QUERY
-POSTHOOK: query: explain extended select ds, count(hr) from TEST2 group by ds
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Edges:
-        Reducer 2 <- Map 1 (SIMPLE_EDGE)
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: test2
-                  Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE
-                  GatherStats: false
-                  Select Operator
-                    expressions: ds (type: string), hr (type: string)
-                    outputColumnNames: ds, hr
-                    Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE
-                    Group By Operator
-                      aggregations: count(hr)
-                      keys: ds (type: string)
-                      mode: hash
-                      outputColumnNames: _col0, _col1
-                      Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: COMPLETE
-                      Reduce Output Operator
-                        key expressions: _col0 (type: string)
-                        null sort order: a
-                        sort order: +
-                        Map-reduce partition columns: _col0 (type: string)
-                        Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: COMPLETE
-                        tag: -1
-                        value expressions: _col1 (type: bigint)
-                        auto parallelism: true
-            Path -> Alias:
-#### A masked pattern was here ####
-            Path -> Partition:
-#### A masked pattern was here ####
-                Partition
-                  base file name: hr=1
-                  input format: org.apache.hadoop.mapred.TextInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  partition values:
-                    ds 1
-                    hr 1
-                  properties:
-                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
-                    bucket_count -1
-                    columns a,b
-                    columns.comments 
-                    columns.types int:double
-#### A masked pattern was here ####
-                    name default.test2
-                    numFiles 0
-                    numRows 0
-                    partition_columns ds/hr
-                    partition_columns.types string:string
-                    rawDataSize 0
-                    serialization.ddl struct test2 { i32 a, double b}
-                    serialization.format 1
-                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    totalSize 0
-#### A masked pattern was here ####
-                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns a,b
-                      columns.comments 
-                      columns.types int:double
-#### A masked pattern was here ####
-                      name default.test2
-                      partition_columns ds/hr
-                      partition_columns.types string:string
-                      serialization.ddl struct test2 { i32 a, double b}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.test2
-                  name: default.test2
-#### A masked pattern was here ####
-                Partition
-                  base file name: hr=2
-                  input format: org.apache.hadoop.mapred.TextInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  partition values:
-                    ds 1
-                    hr 2
-                  properties:
-                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
-                    bucket_count -1
-                    columns a,b
-                    columns.comments 
-                    columns.types int:double
-#### A masked pattern was here ####
-                    name default.test2
-                    numFiles 0
-                    numRows 0
-                    partition_columns ds/hr
-                    partition_columns.types string:string
-                    rawDataSize 0
-                    serialization.ddl struct test2 { i32 a, double b}
-                    serialization.format 1
-                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    totalSize 0
-#### A masked pattern was here ####
-                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns a,b
-                      columns.comments 
-                      columns.types int:double
-#### A masked pattern was here ####
-                      name default.test2
-                      partition_columns ds/hr
-                      partition_columns.types string:string
-                      serialization.ddl struct test2 { i32 a, double b}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.test2
-                  name: default.test2
-#### A masked pattern was here ####
-                Partition
-                  base file name: hr=3
-                  input format: org.apache.hadoop.mapred.TextInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  partition values:
-                    ds 1
-                    hr 3
-                  properties:
-                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
-                    bucket_count -1
-                    columns a,b
-                    columns.comments 
-                    columns.types int:double
-#### A masked pattern was here ####
-                    name default.test2
-                    numFiles 0
-                    numRows 0
-                    partition_columns ds/hr
-                    partition_columns.types string:string
-                    rawDataSize 0
-                    serialization.ddl struct test2 { i32 a, double b}
-                    serialization.format 1
-                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    totalSize 0
-#### A masked pattern was here ####
-                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns a,b
-                      columns.comments 
-                      columns.types int:double
-#### A masked pattern was here ####
-                      name default.test2
-                      partition_columns ds/hr
-                      partition_columns.types string:string
-                      serialization.ddl struct test2 { i32 a, double b}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.test2
-                  name: default.test2
-            Truncated Path -> Alias:
-              /test2/ds=1/hr=1 [test2]
-              /test2/ds=1/hr=2 [test2]
-              /test2/ds=1/hr=3 [test2]
-        Reducer 2 
-            Needs Tagging: false
-            Reduce Operator Tree:
-              Group By Operator
-                aggregations: count(VALUE._col0)
-                keys: KEY._col0 (type: string)
-                mode: mergepartial
-                outputColumnNames: _col0, _col1
-                Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: COMPLETE
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-#### A masked pattern was here ####
-                  NumFilesPerFileSink: 1
-                  Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: COMPLETE
-#### A masked pattern was here ####
-                  table:
-                      input format: org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      properties:
-                        columns _col0,_col1
-                        columns.types string:bigint
-                        escape.delim \
-                        hive.serialization.extend.additional.nesting.levels true
-                        serialization.escape.crlf true
-                        serialization.format 1
-                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                  TotalFiles: 1
-                  GatherStats: false
-                  MultiFileSpray: false
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
-
-PREHOOK: query: select ds, count(hr) from TEST2 group by ds
-PREHOOK: type: QUERY
-PREHOOK: Input: default@test2
-PREHOOK: Input: default@test2@ds=1/hr=1
-PREHOOK: Input: default@test2@ds=1/hr=2
-PREHOOK: Input: default@test2@ds=1/hr=3
-#### A masked pattern was here ####
-POSTHOOK: query: select ds, count(hr) from TEST2 group by ds
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@test2
-POSTHOOK: Input: default@test2@ds=1/hr=1
-POSTHOOK: Input: default@test2@ds=1/hr=2
-POSTHOOK: Input: default@test2@ds=1/hr=3
-#### A masked pattern was here ####
-PREHOOK: query: explain extended select max(ds) from TEST1
-PREHOOK: type: QUERY
-POSTHOOK: query: explain extended select max(ds) from TEST1
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Edges:
-        Reducer 2 <- Map 1 (SIMPLE_EDGE)
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: test1
-                  Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                  GatherStats: false
-                  Select Operator
-                    expressions: ds (type: string)
-                    outputColumnNames: ds
-                    Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                    Group By Operator
-                      aggregations: max(ds)
-                      mode: hash
-                      outputColumnNames: _col0
-                      Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                      Reduce Output Operator
-                        null sort order: 
-                        sort order: 
-                        Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                        tag: -1
-                        value expressions: _col0 (type: string)
-                        auto parallelism: false
-            Path -> Alias:
-              nullscan://null/default.test1/part_ds=1_ [test1]
-              nullscan://null/default.test1/part_ds=2_ [test1]
-            Path -> Partition:
-              nullscan://null/default.test1/part_ds=1_ 
-                Partition
-                  input format: org.apache.hadoop.hive.ql.io.OneNullRowInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  partition values:
-                    ds 1
-                  properties:
-                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
-                    bucket_count -1
-                    columns a,b
-                    columns.comments 
-                    columns.types int:double
-#### A masked pattern was here ####
-                    name default.test1
-                    numFiles 0
-                    numRows 0
-                    partition_columns ds
-                    partition_columns.types string
-                    rawDataSize 0
-                    serialization.ddl struct test1 { i32 a, double b}
-                    serialization.format 1
-                    serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
-                    totalSize 0
-#### A masked pattern was here ####
-                  serde: org.apache.hadoop.hive.serde2.NullStructSerDe
-                
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns a,b
-                      columns.comments 
-                      columns.types int:double
-#### A masked pattern was here ####
-                      name default.test1
-                      partition_columns ds
-                      partition_columns.types string
-                      serialization.ddl struct test1 { i32 a, double b}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.test1
-                  name: default.test1
-              nullscan://null/default.test1/part_ds=2_ 
-                Partition
-                  input format: org.apache.hadoop.hive.ql.io.OneNullRowInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  partition values:
-                    ds 2
-                  properties:
-                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
-                    bucket_count -1
-                    columns a,b
-                    columns.comments 
-                    columns.types int:double
-#### A masked pattern was here ####
-                    name default.test1
-                    numFiles 0
-                    numRows 0
-                    partition_columns ds
-                    partition_columns.types string
-                    rawDataSize 0
-                    serialization.ddl struct test1 { i32 a, double b}
-                    serialization.format 1
-                    serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
-                    totalSize 0
-#### A masked pattern was here ####
-                  serde: org.apache.hadoop.hive.serde2.NullStructSerDe
-                
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns a,b
-                      columns.comments 
-                      columns.types int:double
-#### A masked pattern was here ####
-                      name default.test1
-                      partition_columns ds
-                      partition_columns.types string
-                      serialization.ddl struct test1 { i32 a, double b}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.test1
-                  name: default.test1
-            Truncated Path -> Alias:
-              nullscan://null/default.test1/part_ds=1_ [test1]
-              nullscan://null/default.test1/part_ds=2_ [test1]
-        Reducer 2 
-            Needs Tagging: false
-            Reduce Operator Tree:
-              Group By Operator
-                aggregations: max(VALUE._col0)
-                mode: mergepartial
-                outputColumnNames: _col0
-                Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-#### A masked pattern was here ####
-                  NumFilesPerFileSink: 1
-                  Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE
-#### A masked pattern was here ####
-                  table:
-                      input format: org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      properties:
-                        columns _col0
-                        columns.types string
-                        escape.delim \
-                        hive.serialization.extend.additional.nesting.levels true
-                        serialization.escape.crlf true
-                        serialization.format 1
-                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                  TotalFiles: 1
-                  GatherStats: false
-                  MultiFileSpray: false
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
-
-PREHOOK: query: select max(ds) from TEST1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@test1
-PREHOOK: Input: default@test1@ds=1
-PREHOOK: Input: default@test1@ds=2
-#### A masked pattern was here ####
-POSTHOOK: query: select max(ds) from TEST1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@test1
-POSTHOOK: Input: default@test1@ds=1
-POSTHOOK: Input: default@test1@ds=2
-#### A masked pattern was here ####
-2
-PREHOOK: query: select distinct ds from srcpart
-PREHOOK: type: QUERY
-PREHOOK: Input: default@srcpart
-PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
-PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
-PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
-#### A masked pattern was here ####
-POSTHOOK: query: select distinct ds from srcpart
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@srcpart
-POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
-POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
-POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
-#### A masked pattern was here ####
-2008-04-08
-2008-04-09
-PREHOOK: query: select min(ds),max(ds) from srcpart
-PREHOOK: type: QUERY
-PREHOOK: Input: default@srcpart
-PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
-PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
-PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
-#### A masked pattern was here ####
-POSTHOOK: query: select min(ds),max(ds) from srcpart
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@srcpart
-POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
-POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
-POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
-#### A masked pattern was here ####
-2008-04-08	2008-04-09
-PREHOOK: query: -- HIVE-3594 URI encoding for temporary path
-alter table TEST2 add partition (ds='01:10:10', hr='01')
-PREHOOK: type: ALTERTABLE_ADDPARTS
-PREHOOK: Output: default@test2
-POSTHOOK: query: -- HIVE-3594 URI encoding for temporary path
-alter table TEST2 add partition (ds='01:10:10', hr='01')
-POSTHOOK: type: ALTERTABLE_ADDPARTS
-POSTHOOK: Output: default@test2
-POSTHOOK: Output: default@test2@ds=01%3A10%3A10/hr=01
-PREHOOK: query: alter table TEST2 add partition (ds='01:10:20', hr='02')
-PREHOOK: type: ALTERTABLE_ADDPARTS
-PREHOOK: Output: default@test2
-POSTHOOK: query: alter table TEST2 add partition (ds='01:10:20', hr='02')
-POSTHOOK: type: ALTERTABLE_ADDPARTS
-POSTHOOK: Output: default@test2
-POSTHOOK: Output: default@test2@ds=01%3A10%3A20/hr=02
-PREHOOK: query: explain extended select ds, count(distinct hr) from TEST2 group by ds
-PREHOOK: type: QUERY
-POSTHOOK: query: explain extended select ds, count(distinct hr) from TEST2 group by ds
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Edges:
-        Reducer 2 <- Map 1 (SIMPLE_EDGE)
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: test2
-                  Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE
-                  GatherStats: false
-                  Select Operator
-                    expressions: ds (type: string), hr (type: string)
-                    outputColumnNames: ds, hr
-                    Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE
-                    Group By Operator
-                      aggregations: count(DISTINCT hr)
-                      keys: ds (type: string), hr (type: string)
-                      mode: hash
-                      outputColumnNames: _col0, _col1, _col2
-                      Statistics: Num rows: 1 Data size: 376 Basic stats: COMPLETE Column stats: COMPLETE
-                      Reduce Output Operator
-                        key expressions: _col0 (type: string), _col1 (type: string)
-                        null sort order: aa
-                        sort order: ++
-                        Map-reduce partition columns: _col0 (type: string)
-                        Statistics: Num rows: 1 Data size: 376 Basic stats: COMPLETE Column stats: COMPLETE
-                        tag: -1
-                        auto parallelism: true
-            Path -> Alias:
-              nullscan://null/default.test2/part_ds=01_10_10_hr=01_ [test2]
-              nullscan://null/default.test2/part_ds=01_10_20_hr=02_ [test2]
-              nullscan://null/default.test2/part_ds=1_hr=1_ [test2]
-              nullscan://null/default.test2/part_ds=1_hr=2_ [test2]
-              nullscan://null/default.test2/part_ds=1_hr=3_ [test2]
-            Path -> Partition:
-              nullscan://null/default.test2/part_ds=01_10_10_hr=01_ 
-                Partition
-                  input format: org.apache.hadoop.hive.ql.io.OneNullRowInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  partition values:
-                    ds 01:10:10
-                    hr 01
-                  properties:
-                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
-                    bucket_count -1
-                    columns a,b
-                    columns.comments 
-                    columns.types int:double
-#### A masked pattern was here ####
-                    name default.test2
-                    numFiles 0
-                    numRows 0
-                    partition_columns ds/hr
-                    partition_columns.types string:string
-                    rawDataSize 0
-                    serialization.ddl struct test2 { i32 a, double b}
-                    serialization.format 1
-                    serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
-                    totalSize 0
-#### A masked pattern was here ####
-                  serde: org.apache.hadoop.hive.serde2.NullStructSerDe
-                
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns a,b
-                      columns.comments 
-                      columns.types int:double
-#### A masked pattern was here ####
-                      name default.test2
-                      partition_columns ds/hr
-                      partition_columns.types string:string
-                      serialization.ddl struct test2 { i32 a, double b}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.test2
-                  name: default.test2
-              nullscan://null/default.test2/part_ds=01_10_20_hr=02_ 
-                Partition
-                  input format: org.apache.hadoop.hive.ql.io.OneNullRowInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  partition values:
-                    ds 01:10:20
-                    hr 02
-                  properties:
-                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
-                    bucket_count -1
-                    columns a,b
-                    columns.comments 
-                    columns.types int:double
-#### A masked pattern was here ####
-                    name default.test2
-                    numFiles 0
-                    numRows 0
-                    partition_columns ds/hr
-                    partition_columns.types string:string
-                    rawDataSize 0
-                    serialization.ddl struct test2 { i32 a, double b}
-                    serialization.format 1
-                    serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
-                    totalSize 0
-#### A masked pattern was here ####
-                  serde: org.apache.hadoop.hive.serde2.NullStructSerDe
-                
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns a,b
-                      columns.comments 
-                      columns.types int:double
-#### A masked pattern was here ####
-                      name default.test2
-                      partition_columns ds/hr
-                      partition_columns.types string:string
-                      serialization.ddl struct test2 { i32 a, double b}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.test2
-                  name: default.test2
-              nullscan://null/default.test2/part_ds=1_hr=1_ 
-                Partition
-                  input format: org.apache.hadoop.hive.ql.io.OneNullRowInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  partition values:
-                    ds 1
-                    hr 1
-                  properties:
-                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
-                    bucket_count -1
-                    columns a,b
-                    columns.comments 
-                    columns.types int:double
-#### A masked pattern was here ####
-                    name default.test2
-                    numFiles 0
-                    numRows 0
-                    partition_columns ds/hr
-                    partition_columns.types string:string
-                    rawDataSize 0
-                    serialization.ddl struct test2 { i32 a, double b}
-                    serialization.format 1
-                    serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
-                    totalSize 0
-#### A masked pattern was here ####
-                  serde: org.apache.hadoop.hive.serde2.NullStructSerDe
-                
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns a,b
-                      columns.comments 
-                      columns.types int:double
-#### A masked pattern was here ####
-                      name default.test2
-                      partition_columns ds/hr
-                      partition_columns.types string:string
-                      serialization.ddl struct test2 { i32 a, double b}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.test2
-                  name: default.test2
-              nullscan://null/default.test2/part_ds=1_hr=2_ 
-                Partition
-                  input format: org.apache.hadoop.hive.ql.io.OneNullRowInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  partition values:
-                    ds 1
-                    hr 2
-                  properties:
-                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
-                    bucket_count -1
-                    columns a,b
-                    columns.comments 
-                    columns.types int:double
-#### A masked pattern was here ####
-                    name default.test2
-                    numFiles 0
-                    numRows 0
-                    partition_columns ds/hr
-                    partition_columns.types string:string
-                    rawDataSize 0
-                    serialization.ddl struct test2 { i32 a, double b}
-                    serialization.format 1
-                    serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
-                    totalSize 0
-#### A masked pattern was here ####
-                  serde: org.apache.hadoop.hive.serde2.NullStructSerDe
-                
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns a,b
-                      columns.comments 
-                      columns.types int:double
-#### A masked pattern was here ####
-                      name default.test2
-                      partition_columns ds/hr
-                      partition_columns.types string:string
-                      serialization.ddl struct test2 { i32 a, double b}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.test2
-                  name: default.test2
-              nullscan://null/default.test2/part_ds=1_hr=3_ 
-                Partition
-                  input format: org.apache.hadoop.hive.ql.io.OneNullRowInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  partition values:
-                    ds 1
-                    hr 3
-                  properties:
-                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
-                    bucket_count -1
-                    columns a,b
-                    columns.comments 
-                    columns.types int:double
-#### A masked pattern was here ####
-                    name default.test2
-                    numFiles 0
-                    numRows 0
-                    partition_columns ds/hr
-                    partition_columns.types string:string
-                    rawDataSize 0
-                    serialization.ddl struct test2 { i32 a, double b}
-                    serialization.format 1
-                    serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
-                    totalSize 0
-#### A masked pattern was here ####
-                  serde: org.apache.hadoop.hive.serde2.NullStructSerDe
-                
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns a,b
-                      columns.comments 
-                      columns.types int:double
-#### A masked pattern was here ####
-                      name default.test2
-                      partition_columns ds/hr
-                      partition_columns.types string:string
-                      serialization.ddl struct test2 { i32 a, double b}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.test2
-                  name: default.test2
-            Truncated Path -> Alias:
-              nullscan://null/default.test2/part_ds=01_10_10_hr=01_ [test2]
-              nullscan://null/default.test2/part_ds=01_10_20_hr=02_ [test2]
-              nullscan://null/default.test2/part_ds=1_hr=1_ [test2]
-              nullscan://null/default.test2/part_ds=1_hr=2_ [test2]
-              nullscan://null/default.test2/part_ds=1_hr=3_ [test2]
-        Reducer 2 
-            Needs Tagging: false
-            Reduce Operator Tree:
-              Group By Operator
-                aggregations: count(DISTINCT KEY._col1:0._col0)
-                keys: KEY._col0 (type: string)
-                mode: mergepartial
-                outputColumnNames: _col0, _col1
-                Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: COMPLETE
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-#### A masked pattern was here ####
-                  NumFilesPerFileSink: 1
-                  Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: COMPLETE
-#### A masked pattern was here ####
-                  table:
-                      input format: org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      properties:
-                        columns _col0,_col1
-                        columns.types string:bigint
-                        escape.delim \
-                        hive.serialization.extend.additional.nesting.levels true
-                        serialization.escape.crlf true
-                        serialization.format 1
-                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                  TotalFiles: 1
-                  GatherStats: false
-                  MultiFileSpray: false
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
-
-PREHOOK: query: select ds, count(distinct hr) from TEST2 group by ds
-PREHOOK: type: QUERY
-PREHOOK: Input: default@test2
-PREHOOK: Input: default@test2@ds=01%3A10%3A10/hr=01
-PREHOOK: Input: default@test2@ds=01%3A10%3A20/hr=02
-PREHOOK: Input: d

<TRUNCATED>