You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by pr...@apache.org on 2015/09/15 22:42:19 UTC
[13/51] [partial] hive git commit: HIVE-11776: LLAP: Generate golden
files for all MiniLlapCluster tests (Prasanth Jayachandran)
http://git-wip-us.apache.org/repos/asf/hive/blob/ace87818/ql/src/test/results/clientpositive/llap/mapjoin_mapjoin.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/mapjoin_mapjoin.q.out b/ql/src/test/results/clientpositive/llap/mapjoin_mapjoin.q.out
new file mode 100644
index 0000000..17be306
--- /dev/null
+++ b/ql/src/test/results/clientpositive/llap/mapjoin_mapjoin.q.out
@@ -0,0 +1,825 @@
+PREHOOK: query: -- Since the inputs are small, it should be automatically converted to mapjoin
+
+-- SORT_QUERY_RESULTS
+
+explain extended select srcpart.key from srcpart join src on (srcpart.value=src.value) join src1 on (srcpart.key=src1.key)
+PREHOOK: type: QUERY
+POSTHOOK: query: -- Since the inputs are small, it should be automatically converted to mapjoin
+
+-- SORT_QUERY_RESULTS
+
+explain extended select srcpart.key from srcpart join src on (srcpart.value=src.value) join src1 on (srcpart.key=src1.key)
+POSTHOOK: type: QUERY
+ABSTRACT SYNTAX TREE:
+
+TOK_QUERY
+ TOK_FROM
+ TOK_JOIN
+ TOK_JOIN
+ TOK_TABREF
+ TOK_TABNAME
+ srcpart
+ TOK_TABREF
+ TOK_TABNAME
+ src
+ =
+ .
+ TOK_TABLE_OR_COL
+ srcpart
+ value
+ .
+ TOK_TABLE_OR_COL
+ src
+ value
+ TOK_TABREF
+ TOK_TABNAME
+ src1
+ =
+ .
+ TOK_TABLE_OR_COL
+ srcpart
+ key
+ .
+ TOK_TABLE_OR_COL
+ src1
+ key
+ TOK_INSERT
+ TOK_DESTINATION
+ TOK_DIR
+ TOK_TMP_FILE
+ TOK_SELECT
+ TOK_SELEXPR
+ .
+ TOK_TABLE_OR_COL
+ srcpart
+ key
+
+
+STAGE DEPENDENCIES:
+ Stage-1 is a root stage
+ Stage-0 depends on stages: Stage-1
+
+STAGE PLANS:
+ Stage: Stage-1
+ Tez
+ Edges:
+ Map 1 <- Map 2 (BROADCAST_EDGE), Map 3 (BROADCAST_EDGE)
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: srcpart
+ Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
+ GatherStats: false
+ Filter Operator
+ isSamplingPred: false
+ predicate: (key is not null and value is not null) (type: boolean)
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string), value (type: string)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Map Join Operator
+ condition map:
+ Inner Join 0 to 1
+ Estimated key counts: Map 2 => 13
+ keys:
+ 0 _col0 (type: string)
+ 1 _col0 (type: string)
+ outputColumnNames: _col0, _col1
+ input vertices:
+ 1 Map 2
+ Position of Big Table: 0
+ Statistics: Num rows: 550 Data size: 5843 Basic stats: COMPLETE Column stats: NONE
+ HybridGraceHashJoin: true
+ Map Join Operator
+ condition map:
+ Inner Join 0 to 1
+ Estimated key counts: Map 3 => 250
+ keys:
+ 0 _col1 (type: string)
+ 1 _col0 (type: string)
+ outputColumnNames: _col0
+ input vertices:
+ 1 Map 3
+ Position of Big Table: 0
+ Statistics: Num rows: 605 Data size: 6427 Basic stats: COMPLETE Column stats: NONE
+ HybridGraceHashJoin: true
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+#### A masked pattern was here ####
+ NumFilesPerFileSink: 1
+ Statistics: Num rows: 605 Data size: 6427 Basic stats: COMPLETE Column stats: NONE
+#### A masked pattern was here ####
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ columns _col0
+ columns.types string
+ escape.delim \
+ hive.serialization.extend.additional.nesting.levels true
+ serialization.format 1
+ serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ TotalFiles: 1
+ GatherStats: false
+ MultiFileSpray: false
+ Execution mode: llap
+ Path -> Alias:
+#### A masked pattern was here ####
+ Path -> Partition:
+#### A masked pattern was here ####
+ Partition
+ base file name: hr=11
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ partition values:
+ ds 2008-04-08
+ hr 11
+ properties:
+ COLUMN_STATS_ACCURATE true
+ bucket_count -1
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.srcpart
+ numFiles 1
+ numRows 500
+ partition_columns ds/hr
+ partition_columns.types string:string
+ rawDataSize 5312
+ serialization.ddl struct srcpart { string key, string value}
+ serialization.format 1
+ serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 5812
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ bucket_count -1
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.srcpart
+ partition_columns ds/hr
+ partition_columns.types string:string
+ serialization.ddl struct srcpart { string key, string value}
+ serialization.format 1
+ serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.srcpart
+ name: default.srcpart
+#### A masked pattern was here ####
+ Partition
+ base file name: hr=12
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ partition values:
+ ds 2008-04-08
+ hr 12
+ properties:
+ COLUMN_STATS_ACCURATE true
+ bucket_count -1
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.srcpart
+ numFiles 1
+ numRows 500
+ partition_columns ds/hr
+ partition_columns.types string:string
+ rawDataSize 5312
+ serialization.ddl struct srcpart { string key, string value}
+ serialization.format 1
+ serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 5812
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ bucket_count -1
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.srcpart
+ partition_columns ds/hr
+ partition_columns.types string:string
+ serialization.ddl struct srcpart { string key, string value}
+ serialization.format 1
+ serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.srcpart
+ name: default.srcpart
+#### A masked pattern was here ####
+ Partition
+ base file name: hr=11
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ partition values:
+ ds 2008-04-09
+ hr 11
+ properties:
+ COLUMN_STATS_ACCURATE true
+ bucket_count -1
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.srcpart
+ numFiles 1
+ numRows 500
+ partition_columns ds/hr
+ partition_columns.types string:string
+ rawDataSize 5312
+ serialization.ddl struct srcpart { string key, string value}
+ serialization.format 1
+ serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 5812
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ bucket_count -1
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.srcpart
+ partition_columns ds/hr
+ partition_columns.types string:string
+ serialization.ddl struct srcpart { string key, string value}
+ serialization.format 1
+ serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.srcpart
+ name: default.srcpart
+#### A masked pattern was here ####
+ Partition
+ base file name: hr=12
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ partition values:
+ ds 2008-04-09
+ hr 12
+ properties:
+ COLUMN_STATS_ACCURATE true
+ bucket_count -1
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.srcpart
+ numFiles 1
+ numRows 500
+ partition_columns ds/hr
+ partition_columns.types string:string
+ rawDataSize 5312
+ serialization.ddl struct srcpart { string key, string value}
+ serialization.format 1
+ serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 5812
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ bucket_count -1
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.srcpart
+ partition_columns ds/hr
+ partition_columns.types string:string
+ serialization.ddl struct srcpart { string key, string value}
+ serialization.format 1
+ serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.srcpart
+ name: default.srcpart
+ Truncated Path -> Alias:
+ /srcpart/ds=2008-04-08/hr=11 [srcpart]
+ /srcpart/ds=2008-04-08/hr=12 [srcpart]
+ /srcpart/ds=2008-04-09/hr=11 [srcpart]
+ /srcpart/ds=2008-04-09/hr=12 [srcpart]
+ Map 2
+ Map Operator Tree:
+ TableScan
+ alias: src1
+ Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
+ GatherStats: false
+ Filter Operator
+ isSamplingPred: false
+ predicate: key is not null (type: boolean)
+ Statistics: Num rows: 13 Data size: 99 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string)
+ outputColumnNames: _col0
+ Statistics: Num rows: 13 Data size: 99 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: string)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 13 Data size: 99 Basic stats: COMPLETE Column stats: NONE
+ tag: 1
+ auto parallelism: true
+ Execution mode: llap
+ Path -> Alias:
+#### A masked pattern was here ####
+ Path -> Partition:
+#### A masked pattern was here ####
+ Partition
+ base file name: src1
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ COLUMN_STATS_ACCURATE true
+ bucket_count -1
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.src1
+ numFiles 1
+ numRows 25
+ rawDataSize 191
+ serialization.ddl struct src1 { string key, string value}
+ serialization.format 1
+ serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 216
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ COLUMN_STATS_ACCURATE true
+ bucket_count -1
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.src1
+ numFiles 1
+ numRows 25
+ rawDataSize 191
+ serialization.ddl struct src1 { string key, string value}
+ serialization.format 1
+ serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 216
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src1
+ name: default.src1
+ Truncated Path -> Alias:
+ /src1 [src1]
+ Map 3
+ Map Operator Tree:
+ TableScan
+ alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ GatherStats: false
+ Filter Operator
+ isSamplingPred: false
+ predicate: value is not null (type: boolean)
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: value (type: string)
+ outputColumnNames: _col0
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: string)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ tag: 1
+ auto parallelism: true
+ Execution mode: llap
+ Path -> Alias:
+#### A masked pattern was here ####
+ Path -> Partition:
+#### A masked pattern was here ####
+ Partition
+ base file name: src
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ COLUMN_STATS_ACCURATE true
+ bucket_count -1
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.src
+ numFiles 1
+ numRows 500
+ rawDataSize 5312
+ serialization.ddl struct src { string key, string value}
+ serialization.format 1
+ serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 5812
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ COLUMN_STATS_ACCURATE true
+ bucket_count -1
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.src
+ numFiles 1
+ numRows 500
+ rawDataSize 5312
+ serialization.ddl struct src { string key, string value}
+ serialization.format 1
+ serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 5812
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src
+ name: default.src
+ Truncated Path -> Alias:
+ /src [src]
+
+ Stage: Stage-0
+ Fetch Operator
+ limit: -1
+ Processor Tree:
+ ListSink
+
+PREHOOK: query: explain
+select srcpart.key from srcpart join src on (srcpart.value=src.value) join src1 on (srcpart.key=src1.key) where srcpart.value > 'val_450'
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+select srcpart.key from srcpart join src on (srcpart.value=src.value) join src1 on (srcpart.key=src1.key) where srcpart.value > 'val_450'
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+ Stage-1 is a root stage
+ Stage-0 depends on stages: Stage-1
+
+STAGE PLANS:
+ Stage: Stage-1
+ Tez
+ Edges:
+ Map 1 <- Map 2 (BROADCAST_EDGE), Map 3 (BROADCAST_EDGE)
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: srcpart
+ Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
+ Filter Operator
+ predicate: (((value > 'val_450') and key is not null) and value is not null) (type: boolean)
+ Statistics: Num rows: 167 Data size: 1774 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string), value (type: string)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 167 Data size: 1774 Basic stats: COMPLETE Column stats: NONE
+ Map Join Operator
+ condition map:
+ Inner Join 0 to 1
+ keys:
+ 0 _col0 (type: string)
+ 1 _col0 (type: string)
+ outputColumnNames: _col0, _col1
+ input vertices:
+ 1 Map 2
+ Statistics: Num rows: 183 Data size: 1951 Basic stats: COMPLETE Column stats: NONE
+ HybridGraceHashJoin: true
+ Map Join Operator
+ condition map:
+ Inner Join 0 to 1
+ keys:
+ 0 _col1 (type: string)
+ 1 _col0 (type: string)
+ outputColumnNames: _col0
+ input vertices:
+ 1 Map 3
+ Statistics: Num rows: 201 Data size: 2146 Basic stats: COMPLETE Column stats: NONE
+ HybridGraceHashJoin: true
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 201 Data size: 2146 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ Execution mode: llap
+ Map 2
+ Map Operator Tree:
+ TableScan
+ alias: src1
+ Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
+ Filter Operator
+ predicate: key is not null (type: boolean)
+ Statistics: Num rows: 13 Data size: 99 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string)
+ outputColumnNames: _col0
+ Statistics: Num rows: 13 Data size: 99 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: string)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 13 Data size: 99 Basic stats: COMPLETE Column stats: NONE
+ Execution mode: llap
+ Map 3
+ Map Operator Tree:
+ TableScan
+ alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Filter Operator
+ predicate: (value > 'val_450') (type: boolean)
+ Statistics: Num rows: 166 Data size: 1763 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: value (type: string)
+ outputColumnNames: _col0
+ Statistics: Num rows: 166 Data size: 1763 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: string)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 166 Data size: 1763 Basic stats: COMPLETE Column stats: NONE
+ Execution mode: llap
+
+ Stage: Stage-0
+ Fetch Operator
+ limit: -1
+ Processor Tree:
+ ListSink
+
+PREHOOK: query: explain
+select count(*) from srcpart join src on (srcpart.value=src.value) join src src1 on (srcpart.key=src1.key) group by ds
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+select count(*) from srcpart join src on (srcpart.value=src.value) join src src1 on (srcpart.key=src1.key) group by ds
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+ Stage-1 is a root stage
+ Stage-0 depends on stages: Stage-1
+
+STAGE PLANS:
+ Stage: Stage-1
+ Tez
+ Edges:
+ Map 1 <- Map 3 (BROADCAST_EDGE), Map 4 (BROADCAST_EDGE)
+ Reducer 2 <- Map 1 (SIMPLE_EDGE)
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: srcpart
+ Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
+ Filter Operator
+ predicate: (value is not null and key is not null) (type: boolean)
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string), value (type: string), ds (type: string)
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Map Join Operator
+ condition map:
+ Inner Join 0 to 1
+ keys:
+ 0 _col1 (type: string)
+ 1 _col0 (type: string)
+ outputColumnNames: _col0, _col2
+ input vertices:
+ 1 Map 3
+ Statistics: Num rows: 550 Data size: 5843 Basic stats: COMPLETE Column stats: NONE
+ HybridGraceHashJoin: true
+ Map Join Operator
+ condition map:
+ Inner Join 0 to 1
+ keys:
+ 0 _col0 (type: string)
+ 1 _col0 (type: string)
+ outputColumnNames: _col2
+ input vertices:
+ 1 Map 4
+ Statistics: Num rows: 605 Data size: 6427 Basic stats: COMPLETE Column stats: NONE
+ HybridGraceHashJoin: true
+ Select Operator
+ expressions: _col2 (type: string)
+ outputColumnNames: _col0
+ Statistics: Num rows: 605 Data size: 6427 Basic stats: COMPLETE Column stats: NONE
+ Group By Operator
+ aggregations: count()
+ keys: _col0 (type: string)
+ mode: hash
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 605 Data size: 6427 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: string)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 605 Data size: 6427 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col1 (type: bigint)
+ Execution mode: llap
+ Map 3
+ Map Operator Tree:
+ TableScan
+ alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Filter Operator
+ predicate: value is not null (type: boolean)
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: value (type: string)
+ outputColumnNames: _col0
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: string)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Execution mode: llap
+ Map 4
+ Map Operator Tree:
+ TableScan
+ alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Filter Operator
+ predicate: key is not null (type: boolean)
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string)
+ outputColumnNames: _col0
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: string)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Execution mode: llap
+ Reducer 2
+ Execution mode: uber
+ Reduce Operator Tree:
+ Group By Operator
+ aggregations: count(VALUE._col0)
+ keys: KEY._col0 (type: string)
+ mode: mergepartial
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 302 Data size: 3208 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col1 (type: bigint)
+ outputColumnNames: _col0
+ Statistics: Num rows: 302 Data size: 3208 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 302 Data size: 3208 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+ Stage: Stage-0
+ Fetch Operator
+ limit: -1
+ Processor Tree:
+ ListSink
+
+PREHOOK: query: select srcpart.key from srcpart join src on (srcpart.value=src.value) join src1 on (srcpart.key=src1.key) where srcpart.value > 'val_450'
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Input: default@src1
+PREHOOK: Input: default@srcpart
+PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
+PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
+PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
+PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
+#### A masked pattern was here ####
+POSTHOOK: query: select srcpart.key from srcpart join src on (srcpart.value=src.value) join src1 on (srcpart.key=src1.key) where srcpart.value > 'val_450'
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Input: default@src1
+POSTHOOK: Input: default@srcpart
+POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
+POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
+POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
+POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
+#### A masked pattern was here ####
+66
+66
+66
+66
+98
+98
+98
+98
+98
+98
+98
+98
+98
+98
+98
+98
+98
+98
+98
+98
+PREHOOK: query: select count(*) from srcpart join src on (srcpart.value=src.value) join src src1 on (srcpart.key=src1.key) group by ds
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Input: default@srcpart
+PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
+PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
+PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
+PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
+#### A masked pattern was here ####
+POSTHOOK: query: select count(*) from srcpart join src on (srcpart.value=src.value) join src src1 on (srcpart.key=src1.key) group by ds
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Input: default@srcpart
+POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
+POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
+POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
+POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
+#### A masked pattern was here ####
+5308
+5308
+PREHOOK: query: select srcpart.key from srcpart join src on (srcpart.value=src.value) join src1 on (srcpart.key=src1.key) where srcpart.value > 'val_450'
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Input: default@src1
+PREHOOK: Input: default@srcpart
+PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
+PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
+PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
+PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
+#### A masked pattern was here ####
+POSTHOOK: query: select srcpart.key from srcpart join src on (srcpart.value=src.value) join src1 on (srcpart.key=src1.key) where srcpart.value > 'val_450'
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Input: default@src1
+POSTHOOK: Input: default@srcpart
+POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
+POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
+POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
+POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
+#### A masked pattern was here ####
+66
+66
+66
+66
+98
+98
+98
+98
+98
+98
+98
+98
+98
+98
+98
+98
+98
+98
+98
+98
+PREHOOK: query: select count(*) from srcpart join src on (srcpart.value=src.value) join src src1 on (srcpart.key=src1.key) group by ds
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Input: default@srcpart
+PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
+PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
+PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
+PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
+#### A masked pattern was here ####
+POSTHOOK: query: select count(*) from srcpart join src on (srcpart.value=src.value) join src src1 on (srcpart.key=src1.key) group by ds
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Input: default@srcpart
+POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
+POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
+POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
+POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
+#### A masked pattern was here ####
+5308
+5308
http://git-wip-us.apache.org/repos/asf/hive/blob/ace87818/ql/src/test/results/clientpositive/llap/mapreduce1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/mapreduce1.q.out b/ql/src/test/results/clientpositive/llap/mapreduce1.q.out
new file mode 100644
index 0000000..6eb5ecc
--- /dev/null
+++ b/ql/src/test/results/clientpositive/llap/mapreduce1.q.out
@@ -0,0 +1,621 @@
+PREHOOK: query: CREATE TABLE dest1(key INT, ten INT, one INT, value STRING) STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@dest1
+POSTHOOK: query: CREATE TABLE dest1(key INT, ten INT, one INT, value STRING) STORED AS TEXTFILE
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@dest1
+PREHOOK: query: EXPLAIN
+FROM src
+INSERT OVERWRITE TABLE dest1
+MAP src.key, CAST(src.key / 10 AS INT), CAST(src.key % 10 AS INT), src.value
+USING 'cat' AS (tkey, ten, one, tvalue)
+DISTRIBUTE BY tvalue, tkey
+SORT BY ten, one
+PREHOOK: type: QUERY
+POSTHOOK: query: EXPLAIN
+FROM src
+INSERT OVERWRITE TABLE dest1
+MAP src.key, CAST(src.key / 10 AS INT), CAST(src.key % 10 AS INT), src.value
+USING 'cat' AS (tkey, ten, one, tvalue)
+DISTRIBUTE BY tvalue, tkey
+SORT BY ten, one
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+ Stage-1 is a root stage
+ Stage-2 depends on stages: Stage-1
+ Stage-0 depends on stages: Stage-2
+ Stage-3 depends on stages: Stage-0
+
+STAGE PLANS:
+ Stage: Stage-1
+ Tez
+ Edges:
+ Reducer 2 <- Map 1 (SIMPLE_EDGE)
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string), UDFToInteger((key / 10)) (type: int), UDFToInteger((key % 10)) (type: int), value (type: string)
+ outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Transform Operator
+ command: cat
+ output info:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col1 (type: string), _col2 (type: string)
+ sort order: ++
+ Map-reduce partition columns: _col3 (type: string), _col0 (type: string)
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string)
+ Reducer 2
+ Execution mode: uber
+ Reduce Operator Tree:
+ Select Operator
+ expressions: UDFToInteger(VALUE._col0) (type: int), UDFToInteger(VALUE._col1) (type: int), UDFToInteger(VALUE._col2) (type: int), VALUE._col3 (type: string)
+ outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.dest1
+
+ Stage: Stage-2
+ Dependency Collection
+
+ Stage: Stage-0
+ Move Operator
+ tables:
+ replace: true
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.dest1
+
+ Stage: Stage-3
+ Stats-Aggr Operator
+
+PREHOOK: query: FROM src
+INSERT OVERWRITE TABLE dest1
+MAP src.key, CAST(src.key / 10 AS INT), CAST(src.key % 10 AS INT), src.value
+USING 'cat' AS (tkey, ten, one, tvalue)
+DISTRIBUTE BY tvalue, tkey
+SORT BY ten, one
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@dest1
+POSTHOOK: query: FROM src
+INSERT OVERWRITE TABLE dest1
+MAP src.key, CAST(src.key / 10 AS INT), CAST(src.key % 10 AS INT), src.value
+USING 'cat' AS (tkey, ten, one, tvalue)
+DISTRIBUTE BY tvalue, tkey
+SORT BY ten, one
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@dest1
+POSTHOOK: Lineage: dest1.key SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: dest1.one SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: dest1.ten SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: dest1.value SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
+PREHOOK: query: SELECT dest1.* FROM dest1
+PREHOOK: type: QUERY
+PREHOOK: Input: default@dest1
+#### A masked pattern was here ####
+POSTHOOK: query: SELECT dest1.* FROM dest1
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@dest1
+#### A masked pattern was here ####
+0 0 0 val_0
+0 0 0 val_0
+0 0 0 val_0
+2 0 2 val_2
+4 0 4 val_4
+5 0 5 val_5
+5 0 5 val_5
+5 0 5 val_5
+8 0 8 val_8
+9 0 9 val_9
+10 1 0 val_10
+11 1 1 val_11
+12 1 2 val_12
+12 1 2 val_12
+15 1 5 val_15
+15 1 5 val_15
+17 1 7 val_17
+18 1 8 val_18
+18 1 8 val_18
+19 1 9 val_19
+100 10 0 val_100
+100 10 0 val_100
+103 10 3 val_103
+103 10 3 val_103
+104 10 4 val_104
+104 10 4 val_104
+105 10 5 val_105
+111 11 1 val_111
+113 11 3 val_113
+113 11 3 val_113
+114 11 4 val_114
+116 11 6 val_116
+118 11 8 val_118
+118 11 8 val_118
+119 11 9 val_119
+119 11 9 val_119
+119 11 9 val_119
+120 12 0 val_120
+120 12 0 val_120
+125 12 5 val_125
+125 12 5 val_125
+126 12 6 val_126
+128 12 8 val_128
+128 12 8 val_128
+128 12 8 val_128
+129 12 9 val_129
+129 12 9 val_129
+131 13 1 val_131
+133 13 3 val_133
+134 13 4 val_134
+134 13 4 val_134
+136 13 6 val_136
+137 13 7 val_137
+137 13 7 val_137
+138 13 8 val_138
+138 13 8 val_138
+138 13 8 val_138
+138 13 8 val_138
+143 14 3 val_143
+145 14 5 val_145
+146 14 6 val_146
+146 14 6 val_146
+149 14 9 val_149
+149 14 9 val_149
+150 15 0 val_150
+152 15 2 val_152
+152 15 2 val_152
+153 15 3 val_153
+155 15 5 val_155
+156 15 6 val_156
+157 15 7 val_157
+158 15 8 val_158
+160 16 0 val_160
+162 16 2 val_162
+163 16 3 val_163
+164 16 4 val_164
+164 16 4 val_164
+165 16 5 val_165
+165 16 5 val_165
+166 16 6 val_166
+167 16 7 val_167
+167 16 7 val_167
+167 16 7 val_167
+168 16 8 val_168
+169 16 9 val_169
+169 16 9 val_169
+169 16 9 val_169
+169 16 9 val_169
+170 17 0 val_170
+172 17 2 val_172
+172 17 2 val_172
+174 17 4 val_174
+174 17 4 val_174
+175 17 5 val_175
+175 17 5 val_175
+176 17 6 val_176
+176 17 6 val_176
+177 17 7 val_177
+178 17 8 val_178
+179 17 9 val_179
+179 17 9 val_179
+180 18 0 val_180
+181 18 1 val_181
+183 18 3 val_183
+186 18 6 val_186
+187 18 7 val_187
+187 18 7 val_187
+187 18 7 val_187
+189 18 9 val_189
+190 19 0 val_190
+191 19 1 val_191
+191 19 1 val_191
+192 19 2 val_192
+193 19 3 val_193
+193 19 3 val_193
+193 19 3 val_193
+194 19 4 val_194
+195 19 5 val_195
+195 19 5 val_195
+196 19 6 val_196
+197 19 7 val_197
+197 19 7 val_197
+199 19 9 val_199
+199 19 9 val_199
+199 19 9 val_199
+20 2 0 val_20
+24 2 4 val_24
+24 2 4 val_24
+26 2 6 val_26
+26 2 6 val_26
+27 2 7 val_27
+28 2 8 val_28
+200 20 0 val_200
+200 20 0 val_200
+201 20 1 val_201
+202 20 2 val_202
+203 20 3 val_203
+203 20 3 val_203
+205 20 5 val_205
+205 20 5 val_205
+207 20 7 val_207
+207 20 7 val_207
+208 20 8 val_208
+208 20 8 val_208
+208 20 8 val_208
+209 20 9 val_209
+209 20 9 val_209
+213 21 3 val_213
+213 21 3 val_213
+214 21 4 val_214
+216 21 6 val_216
+216 21 6 val_216
+217 21 7 val_217
+217 21 7 val_217
+218 21 8 val_218
+219 21 9 val_219
+219 21 9 val_219
+221 22 1 val_221
+221 22 1 val_221
+222 22 2 val_222
+223 22 3 val_223
+223 22 3 val_223
+224 22 4 val_224
+224 22 4 val_224
+226 22 6 val_226
+228 22 8 val_228
+229 22 9 val_229
+229 22 9 val_229
+230 23 0 val_230
+230 23 0 val_230
+230 23 0 val_230
+230 23 0 val_230
+230 23 0 val_230
+233 23 3 val_233
+233 23 3 val_233
+235 23 5 val_235
+237 23 7 val_237
+237 23 7 val_237
+238 23 8 val_238
+238 23 8 val_238
+239 23 9 val_239
+239 23 9 val_239
+241 24 1 val_241
+242 24 2 val_242
+242 24 2 val_242
+244 24 4 val_244
+247 24 7 val_247
+248 24 8 val_248
+249 24 9 val_249
+252 25 2 val_252
+255 25 5 val_255
+255 25 5 val_255
+256 25 6 val_256
+256 25 6 val_256
+257 25 7 val_257
+258 25 8 val_258
+260 26 0 val_260
+262 26 2 val_262
+263 26 3 val_263
+265 26 5 val_265
+265 26 5 val_265
+266 26 6 val_266
+272 27 2 val_272
+272 27 2 val_272
+273 27 3 val_273
+273 27 3 val_273
+273 27 3 val_273
+274 27 4 val_274
+275 27 5 val_275
+277 27 7 val_277
+277 27 7 val_277
+277 27 7 val_277
+277 27 7 val_277
+278 27 8 val_278
+278 27 8 val_278
+280 28 0 val_280
+280 28 0 val_280
+281 28 1 val_281
+281 28 1 val_281
+282 28 2 val_282
+282 28 2 val_282
+283 28 3 val_283
+284 28 4 val_284
+285 28 5 val_285
+286 28 6 val_286
+287 28 7 val_287
+288 28 8 val_288
+288 28 8 val_288
+289 28 9 val_289
+291 29 1 val_291
+292 29 2 val_292
+296 29 6 val_296
+298 29 8 val_298
+298 29 8 val_298
+298 29 8 val_298
+30 3 0 val_30
+33 3 3 val_33
+34 3 4 val_34
+35 3 5 val_35
+35 3 5 val_35
+35 3 5 val_35
+37 3 7 val_37
+37 3 7 val_37
+302 30 2 val_302
+305 30 5 val_305
+306 30 6 val_306
+307 30 7 val_307
+307 30 7 val_307
+308 30 8 val_308
+309 30 9 val_309
+309 30 9 val_309
+310 31 0 val_310
+311 31 1 val_311
+311 31 1 val_311
+311 31 1 val_311
+315 31 5 val_315
+316 31 6 val_316
+316 31 6 val_316
+316 31 6 val_316
+317 31 7 val_317
+317 31 7 val_317
+318 31 8 val_318
+318 31 8 val_318
+318 31 8 val_318
+321 32 1 val_321
+321 32 1 val_321
+322 32 2 val_322
+322 32 2 val_322
+323 32 3 val_323
+325 32 5 val_325
+325 32 5 val_325
+327 32 7 val_327
+327 32 7 val_327
+327 32 7 val_327
+331 33 1 val_331
+331 33 1 val_331
+332 33 2 val_332
+333 33 3 val_333
+333 33 3 val_333
+335 33 5 val_335
+336 33 6 val_336
+338 33 8 val_338
+339 33 9 val_339
+341 34 1 val_341
+342 34 2 val_342
+342 34 2 val_342
+344 34 4 val_344
+344 34 4 val_344
+345 34 5 val_345
+348 34 8 val_348
+348 34 8 val_348
+348 34 8 val_348
+348 34 8 val_348
+348 34 8 val_348
+351 35 1 val_351
+353 35 3 val_353
+353 35 3 val_353
+356 35 6 val_356
+360 36 0 val_360
+362 36 2 val_362
+364 36 4 val_364
+365 36 5 val_365
+366 36 6 val_366
+367 36 7 val_367
+367 36 7 val_367
+368 36 8 val_368
+369 36 9 val_369
+369 36 9 val_369
+369 36 9 val_369
+373 37 3 val_373
+374 37 4 val_374
+375 37 5 val_375
+377 37 7 val_377
+378 37 8 val_378
+379 37 9 val_379
+382 38 2 val_382
+382 38 2 val_382
+384 38 4 val_384
+384 38 4 val_384
+384 38 4 val_384
+386 38 6 val_386
+389 38 9 val_389
+392 39 2 val_392
+393 39 3 val_393
+394 39 4 val_394
+395 39 5 val_395
+395 39 5 val_395
+396 39 6 val_396
+396 39 6 val_396
+396 39 6 val_396
+397 39 7 val_397
+397 39 7 val_397
+399 39 9 val_399
+399 39 9 val_399
+41 4 1 val_41
+42 4 2 val_42
+42 4 2 val_42
+43 4 3 val_43
+44 4 4 val_44
+47 4 7 val_47
+400 40 0 val_400
+401 40 1 val_401
+401 40 1 val_401
+401 40 1 val_401
+401 40 1 val_401
+401 40 1 val_401
+402 40 2 val_402
+403 40 3 val_403
+403 40 3 val_403
+403 40 3 val_403
+404 40 4 val_404
+404 40 4 val_404
+406 40 6 val_406
+406 40 6 val_406
+406 40 6 val_406
+406 40 6 val_406
+407 40 7 val_407
+409 40 9 val_409
+409 40 9 val_409
+409 40 9 val_409
+411 41 1 val_411
+413 41 3 val_413
+413 41 3 val_413
+414 41 4 val_414
+414 41 4 val_414
+417 41 7 val_417
+417 41 7 val_417
+417 41 7 val_417
+418 41 8 val_418
+419 41 9 val_419
+421 42 1 val_421
+424 42 4 val_424
+424 42 4 val_424
+427 42 7 val_427
+429 42 9 val_429
+429 42 9 val_429
+430 43 0 val_430
+430 43 0 val_430
+430 43 0 val_430
+431 43 1 val_431
+431 43 1 val_431
+431 43 1 val_431
+432 43 2 val_432
+435 43 5 val_435
+436 43 6 val_436
+437 43 7 val_437
+438 43 8 val_438
+438 43 8 val_438
+438 43 8 val_438
+439 43 9 val_439
+439 43 9 val_439
+443 44 3 val_443
+444 44 4 val_444
+446 44 6 val_446
+448 44 8 val_448
+449 44 9 val_449
+452 45 2 val_452
+453 45 3 val_453
+454 45 4 val_454
+454 45 4 val_454
+454 45 4 val_454
+455 45 5 val_455
+457 45 7 val_457
+458 45 8 val_458
+458 45 8 val_458
+459 45 9 val_459
+459 45 9 val_459
+460 46 0 val_460
+462 46 2 val_462
+462 46 2 val_462
+463 46 3 val_463
+463 46 3 val_463
+466 46 6 val_466
+466 46 6 val_466
+466 46 6 val_466
+467 46 7 val_467
+468 46 8 val_468
+468 46 8 val_468
+468 46 8 val_468
+468 46 8 val_468
+469 46 9 val_469
+469 46 9 val_469
+469 46 9 val_469
+469 46 9 val_469
+469 46 9 val_469
+470 47 0 val_470
+472 47 2 val_472
+475 47 5 val_475
+477 47 7 val_477
+478 47 8 val_478
+478 47 8 val_478
+479 47 9 val_479
+480 48 0 val_480
+480 48 0 val_480
+480 48 0 val_480
+481 48 1 val_481
+482 48 2 val_482
+483 48 3 val_483
+484 48 4 val_484
+485 48 5 val_485
+487 48 7 val_487
+489 48 9 val_489
+489 48 9 val_489
+489 48 9 val_489
+489 48 9 val_489
+490 49 0 val_490
+491 49 1 val_491
+492 49 2 val_492
+492 49 2 val_492
+493 49 3 val_493
+494 49 4 val_494
+495 49 5 val_495
+496 49 6 val_496
+497 49 7 val_497
+498 49 8 val_498
+498 49 8 val_498
+498 49 8 val_498
+51 5 1 val_51
+51 5 1 val_51
+53 5 3 val_53
+54 5 4 val_54
+57 5 7 val_57
+58 5 8 val_58
+58 5 8 val_58
+64 6 4 val_64
+65 6 5 val_65
+66 6 6 val_66
+67 6 7 val_67
+67 6 7 val_67
+69 6 9 val_69
+70 7 0 val_70
+70 7 0 val_70
+70 7 0 val_70
+72 7 2 val_72
+72 7 2 val_72
+74 7 4 val_74
+76 7 6 val_76
+76 7 6 val_76
+77 7 7 val_77
+78 7 8 val_78
+80 8 0 val_80
+82 8 2 val_82
+83 8 3 val_83
+83 8 3 val_83
+84 8 4 val_84
+84 8 4 val_84
+85 8 5 val_85
+86 8 6 val_86
+87 8 7 val_87
+90 9 0 val_90
+90 9 0 val_90
+90 9 0 val_90
+92 9 2 val_92
+95 9 5 val_95
+95 9 5 val_95
+96 9 6 val_96
+97 9 7 val_97
+97 9 7 val_97
+98 9 8 val_98
+98 9 8 val_98
http://git-wip-us.apache.org/repos/asf/hive/blob/ace87818/ql/src/test/results/clientpositive/llap/mapreduce2.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/mapreduce2.q.out b/ql/src/test/results/clientpositive/llap/mapreduce2.q.out
new file mode 100644
index 0000000..2b18150
--- /dev/null
+++ b/ql/src/test/results/clientpositive/llap/mapreduce2.q.out
@@ -0,0 +1,616 @@
+PREHOOK: query: CREATE TABLE dest1(key INT, ten INT, one INT, value STRING) STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@dest1
+POSTHOOK: query: CREATE TABLE dest1(key INT, ten INT, one INT, value STRING) STORED AS TEXTFILE
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@dest1
+PREHOOK: query: EXPLAIN
+FROM src
+INSERT OVERWRITE TABLE dest1
+MAP src.key, CAST(src.key / 10 AS INT), CAST(src.key % 10 AS INT), src.value
+USING 'cat' AS (tkey, ten, one, tvalue)
+DISTRIBUTE BY tvalue, tkey
+PREHOOK: type: QUERY
+POSTHOOK: query: EXPLAIN
+FROM src
+INSERT OVERWRITE TABLE dest1
+MAP src.key, CAST(src.key / 10 AS INT), CAST(src.key % 10 AS INT), src.value
+USING 'cat' AS (tkey, ten, one, tvalue)
+DISTRIBUTE BY tvalue, tkey
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+ Stage-1 is a root stage
+ Stage-2 depends on stages: Stage-1
+ Stage-0 depends on stages: Stage-2
+ Stage-3 depends on stages: Stage-0
+
+STAGE PLANS:
+ Stage: Stage-1
+ Tez
+ Edges:
+ Reducer 2 <- Map 1 (SIMPLE_EDGE)
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string), UDFToInteger((key / 10)) (type: int), UDFToInteger((key % 10)) (type: int), value (type: string)
+ outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Transform Operator
+ command: cat
+ output info:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ sort order:
+ Map-reduce partition columns: _col3 (type: string), _col0 (type: string)
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string)
+ Reducer 2
+ Execution mode: uber
+ Reduce Operator Tree:
+ Select Operator
+ expressions: UDFToInteger(VALUE._col0) (type: int), UDFToInteger(VALUE._col1) (type: int), UDFToInteger(VALUE._col2) (type: int), VALUE._col3 (type: string)
+ outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.dest1
+
+ Stage: Stage-2
+ Dependency Collection
+
+ Stage: Stage-0
+ Move Operator
+ tables:
+ replace: true
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.dest1
+
+ Stage: Stage-3
+ Stats-Aggr Operator
+
+PREHOOK: query: FROM src
+INSERT OVERWRITE TABLE dest1
+MAP src.key, CAST(src.key / 10 AS INT), CAST(src.key % 10 AS INT), src.value
+USING 'cat' AS (tkey, ten, one, tvalue)
+DISTRIBUTE BY tvalue, tkey
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@dest1
+POSTHOOK: query: FROM src
+INSERT OVERWRITE TABLE dest1
+MAP src.key, CAST(src.key / 10 AS INT), CAST(src.key % 10 AS INT), src.value
+USING 'cat' AS (tkey, ten, one, tvalue)
+DISTRIBUTE BY tvalue, tkey
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@dest1
+POSTHOOK: Lineage: dest1.key SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: dest1.one SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: dest1.ten SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: dest1.value SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
+PREHOOK: query: SELECT * FROM (SELECT dest1.* FROM dest1 DISTRIBUTE BY key SORT BY key, ten, one, value) T ORDER BY key
+PREHOOK: type: QUERY
+PREHOOK: Input: default@dest1
+#### A masked pattern was here ####
+POSTHOOK: query: SELECT * FROM (SELECT dest1.* FROM dest1 DISTRIBUTE BY key SORT BY key, ten, one, value) T ORDER BY key
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@dest1
+#### A masked pattern was here ####
+0 0 0 val_0
+0 0 0 val_0
+0 0 0 val_0
+2 0 2 val_2
+4 0 4 val_4
+5 0 5 val_5
+5 0 5 val_5
+5 0 5 val_5
+8 0 8 val_8
+9 0 9 val_9
+10 1 0 val_10
+11 1 1 val_11
+12 1 2 val_12
+12 1 2 val_12
+15 1 5 val_15
+15 1 5 val_15
+17 1 7 val_17
+18 1 8 val_18
+18 1 8 val_18
+19 1 9 val_19
+20 2 0 val_20
+24 2 4 val_24
+24 2 4 val_24
+26 2 6 val_26
+26 2 6 val_26
+27 2 7 val_27
+28 2 8 val_28
+30 3 0 val_30
+33 3 3 val_33
+34 3 4 val_34
+35 3 5 val_35
+35 3 5 val_35
+35 3 5 val_35
+37 3 7 val_37
+37 3 7 val_37
+41 4 1 val_41
+42 4 2 val_42
+42 4 2 val_42
+43 4 3 val_43
+44 4 4 val_44
+47 4 7 val_47
+51 5 1 val_51
+51 5 1 val_51
+53 5 3 val_53
+54 5 4 val_54
+57 5 7 val_57
+58 5 8 val_58
+58 5 8 val_58
+64 6 4 val_64
+65 6 5 val_65
+66 6 6 val_66
+67 6 7 val_67
+67 6 7 val_67
+69 6 9 val_69
+70 7 0 val_70
+70 7 0 val_70
+70 7 0 val_70
+72 7 2 val_72
+72 7 2 val_72
+74 7 4 val_74
+76 7 6 val_76
+76 7 6 val_76
+77 7 7 val_77
+78 7 8 val_78
+80 8 0 val_80
+82 8 2 val_82
+83 8 3 val_83
+83 8 3 val_83
+84 8 4 val_84
+84 8 4 val_84
+85 8 5 val_85
+86 8 6 val_86
+87 8 7 val_87
+90 9 0 val_90
+90 9 0 val_90
+90 9 0 val_90
+92 9 2 val_92
+95 9 5 val_95
+95 9 5 val_95
+96 9 6 val_96
+97 9 7 val_97
+97 9 7 val_97
+98 9 8 val_98
+98 9 8 val_98
+100 10 0 val_100
+100 10 0 val_100
+103 10 3 val_103
+103 10 3 val_103
+104 10 4 val_104
+104 10 4 val_104
+105 10 5 val_105
+111 11 1 val_111
+113 11 3 val_113
+113 11 3 val_113
+114 11 4 val_114
+116 11 6 val_116
+118 11 8 val_118
+118 11 8 val_118
+119 11 9 val_119
+119 11 9 val_119
+119 11 9 val_119
+120 12 0 val_120
+120 12 0 val_120
+125 12 5 val_125
+125 12 5 val_125
+126 12 6 val_126
+128 12 8 val_128
+128 12 8 val_128
+128 12 8 val_128
+129 12 9 val_129
+129 12 9 val_129
+131 13 1 val_131
+133 13 3 val_133
+134 13 4 val_134
+134 13 4 val_134
+136 13 6 val_136
+137 13 7 val_137
+137 13 7 val_137
+138 13 8 val_138
+138 13 8 val_138
+138 13 8 val_138
+138 13 8 val_138
+143 14 3 val_143
+145 14 5 val_145
+146 14 6 val_146
+146 14 6 val_146
+149 14 9 val_149
+149 14 9 val_149
+150 15 0 val_150
+152 15 2 val_152
+152 15 2 val_152
+153 15 3 val_153
+155 15 5 val_155
+156 15 6 val_156
+157 15 7 val_157
+158 15 8 val_158
+160 16 0 val_160
+162 16 2 val_162
+163 16 3 val_163
+164 16 4 val_164
+164 16 4 val_164
+165 16 5 val_165
+165 16 5 val_165
+166 16 6 val_166
+167 16 7 val_167
+167 16 7 val_167
+167 16 7 val_167
+168 16 8 val_168
+169 16 9 val_169
+169 16 9 val_169
+169 16 9 val_169
+169 16 9 val_169
+170 17 0 val_170
+172 17 2 val_172
+172 17 2 val_172
+174 17 4 val_174
+174 17 4 val_174
+175 17 5 val_175
+175 17 5 val_175
+176 17 6 val_176
+176 17 6 val_176
+177 17 7 val_177
+178 17 8 val_178
+179 17 9 val_179
+179 17 9 val_179
+180 18 0 val_180
+181 18 1 val_181
+183 18 3 val_183
+186 18 6 val_186
+187 18 7 val_187
+187 18 7 val_187
+187 18 7 val_187
+189 18 9 val_189
+190 19 0 val_190
+191 19 1 val_191
+191 19 1 val_191
+192 19 2 val_192
+193 19 3 val_193
+193 19 3 val_193
+193 19 3 val_193
+194 19 4 val_194
+195 19 5 val_195
+195 19 5 val_195
+196 19 6 val_196
+197 19 7 val_197
+197 19 7 val_197
+199 19 9 val_199
+199 19 9 val_199
+199 19 9 val_199
+200 20 0 val_200
+200 20 0 val_200
+201 20 1 val_201
+202 20 2 val_202
+203 20 3 val_203
+203 20 3 val_203
+205 20 5 val_205
+205 20 5 val_205
+207 20 7 val_207
+207 20 7 val_207
+208 20 8 val_208
+208 20 8 val_208
+208 20 8 val_208
+209 20 9 val_209
+209 20 9 val_209
+213 21 3 val_213
+213 21 3 val_213
+214 21 4 val_214
+216 21 6 val_216
+216 21 6 val_216
+217 21 7 val_217
+217 21 7 val_217
+218 21 8 val_218
+219 21 9 val_219
+219 21 9 val_219
+221 22 1 val_221
+221 22 1 val_221
+222 22 2 val_222
+223 22 3 val_223
+223 22 3 val_223
+224 22 4 val_224
+224 22 4 val_224
+226 22 6 val_226
+228 22 8 val_228
+229 22 9 val_229
+229 22 9 val_229
+230 23 0 val_230
+230 23 0 val_230
+230 23 0 val_230
+230 23 0 val_230
+230 23 0 val_230
+233 23 3 val_233
+233 23 3 val_233
+235 23 5 val_235
+237 23 7 val_237
+237 23 7 val_237
+238 23 8 val_238
+238 23 8 val_238
+239 23 9 val_239
+239 23 9 val_239
+241 24 1 val_241
+242 24 2 val_242
+242 24 2 val_242
+244 24 4 val_244
+247 24 7 val_247
+248 24 8 val_248
+249 24 9 val_249
+252 25 2 val_252
+255 25 5 val_255
+255 25 5 val_255
+256 25 6 val_256
+256 25 6 val_256
+257 25 7 val_257
+258 25 8 val_258
+260 26 0 val_260
+262 26 2 val_262
+263 26 3 val_263
+265 26 5 val_265
+265 26 5 val_265
+266 26 6 val_266
+272 27 2 val_272
+272 27 2 val_272
+273 27 3 val_273
+273 27 3 val_273
+273 27 3 val_273
+274 27 4 val_274
+275 27 5 val_275
+277 27 7 val_277
+277 27 7 val_277
+277 27 7 val_277
+277 27 7 val_277
+278 27 8 val_278
+278 27 8 val_278
+280 28 0 val_280
+280 28 0 val_280
+281 28 1 val_281
+281 28 1 val_281
+282 28 2 val_282
+282 28 2 val_282
+283 28 3 val_283
+284 28 4 val_284
+285 28 5 val_285
+286 28 6 val_286
+287 28 7 val_287
+288 28 8 val_288
+288 28 8 val_288
+289 28 9 val_289
+291 29 1 val_291
+292 29 2 val_292
+296 29 6 val_296
+298 29 8 val_298
+298 29 8 val_298
+298 29 8 val_298
+302 30 2 val_302
+305 30 5 val_305
+306 30 6 val_306
+307 30 7 val_307
+307 30 7 val_307
+308 30 8 val_308
+309 30 9 val_309
+309 30 9 val_309
+310 31 0 val_310
+311 31 1 val_311
+311 31 1 val_311
+311 31 1 val_311
+315 31 5 val_315
+316 31 6 val_316
+316 31 6 val_316
+316 31 6 val_316
+317 31 7 val_317
+317 31 7 val_317
+318 31 8 val_318
+318 31 8 val_318
+318 31 8 val_318
+321 32 1 val_321
+321 32 1 val_321
+322 32 2 val_322
+322 32 2 val_322
+323 32 3 val_323
+325 32 5 val_325
+325 32 5 val_325
+327 32 7 val_327
+327 32 7 val_327
+327 32 7 val_327
+331 33 1 val_331
+331 33 1 val_331
+332 33 2 val_332
+333 33 3 val_333
+333 33 3 val_333
+335 33 5 val_335
+336 33 6 val_336
+338 33 8 val_338
+339 33 9 val_339
+341 34 1 val_341
+342 34 2 val_342
+342 34 2 val_342
+344 34 4 val_344
+344 34 4 val_344
+345 34 5 val_345
+348 34 8 val_348
+348 34 8 val_348
+348 34 8 val_348
+348 34 8 val_348
+348 34 8 val_348
+351 35 1 val_351
+353 35 3 val_353
+353 35 3 val_353
+356 35 6 val_356
+360 36 0 val_360
+362 36 2 val_362
+364 36 4 val_364
+365 36 5 val_365
+366 36 6 val_366
+367 36 7 val_367
+367 36 7 val_367
+368 36 8 val_368
+369 36 9 val_369
+369 36 9 val_369
+369 36 9 val_369
+373 37 3 val_373
+374 37 4 val_374
+375 37 5 val_375
+377 37 7 val_377
+378 37 8 val_378
+379 37 9 val_379
+382 38 2 val_382
+382 38 2 val_382
+384 38 4 val_384
+384 38 4 val_384
+384 38 4 val_384
+386 38 6 val_386
+389 38 9 val_389
+392 39 2 val_392
+393 39 3 val_393
+394 39 4 val_394
+395 39 5 val_395
+395 39 5 val_395
+396 39 6 val_396
+396 39 6 val_396
+396 39 6 val_396
+397 39 7 val_397
+397 39 7 val_397
+399 39 9 val_399
+399 39 9 val_399
+400 40 0 val_400
+401 40 1 val_401
+401 40 1 val_401
+401 40 1 val_401
+401 40 1 val_401
+401 40 1 val_401
+402 40 2 val_402
+403 40 3 val_403
+403 40 3 val_403
+403 40 3 val_403
+404 40 4 val_404
+404 40 4 val_404
+406 40 6 val_406
+406 40 6 val_406
+406 40 6 val_406
+406 40 6 val_406
+407 40 7 val_407
+409 40 9 val_409
+409 40 9 val_409
+409 40 9 val_409
+411 41 1 val_411
+413 41 3 val_413
+413 41 3 val_413
+414 41 4 val_414
+414 41 4 val_414
+417 41 7 val_417
+417 41 7 val_417
+417 41 7 val_417
+418 41 8 val_418
+419 41 9 val_419
+421 42 1 val_421
+424 42 4 val_424
+424 42 4 val_424
+427 42 7 val_427
+429 42 9 val_429
+429 42 9 val_429
+430 43 0 val_430
+430 43 0 val_430
+430 43 0 val_430
+431 43 1 val_431
+431 43 1 val_431
+431 43 1 val_431
+432 43 2 val_432
+435 43 5 val_435
+436 43 6 val_436
+437 43 7 val_437
+438 43 8 val_438
+438 43 8 val_438
+438 43 8 val_438
+439 43 9 val_439
+439 43 9 val_439
+443 44 3 val_443
+444 44 4 val_444
+446 44 6 val_446
+448 44 8 val_448
+449 44 9 val_449
+452 45 2 val_452
+453 45 3 val_453
+454 45 4 val_454
+454 45 4 val_454
+454 45 4 val_454
+455 45 5 val_455
+457 45 7 val_457
+458 45 8 val_458
+458 45 8 val_458
+459 45 9 val_459
+459 45 9 val_459
+460 46 0 val_460
+462 46 2 val_462
+462 46 2 val_462
+463 46 3 val_463
+463 46 3 val_463
+466 46 6 val_466
+466 46 6 val_466
+466 46 6 val_466
+467 46 7 val_467
+468 46 8 val_468
+468 46 8 val_468
+468 46 8 val_468
+468 46 8 val_468
+469 46 9 val_469
+469 46 9 val_469
+469 46 9 val_469
+469 46 9 val_469
+469 46 9 val_469
+470 47 0 val_470
+472 47 2 val_472
+475 47 5 val_475
+477 47 7 val_477
+478 47 8 val_478
+478 47 8 val_478
+479 47 9 val_479
+480 48 0 val_480
+480 48 0 val_480
+480 48 0 val_480
+481 48 1 val_481
+482 48 2 val_482
+483 48 3 val_483
+484 48 4 val_484
+485 48 5 val_485
+487 48 7 val_487
+489 48 9 val_489
+489 48 9 val_489
+489 48 9 val_489
+489 48 9 val_489
+490 49 0 val_490
+491 49 1 val_491
+492 49 2 val_492
+492 49 2 val_492
+493 49 3 val_493
+494 49 4 val_494
+495 49 5 val_495
+496 49 6 val_496
+497 49 7 val_497
+498 49 8 val_498
+498 49 8 val_498
+498 49 8 val_498
http://git-wip-us.apache.org/repos/asf/hive/blob/ace87818/ql/src/test/results/clientpositive/llap/merge1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/merge1.q.out b/ql/src/test/results/clientpositive/llap/merge1.q.out
new file mode 100644
index 0000000..ce3cd2c
--- /dev/null
+++ b/ql/src/test/results/clientpositive/llap/merge1.q.out
@@ -0,0 +1,596 @@
+PREHOOK: query: -- SORT_QUERY_RESULTS
+
+create table dest1(key int, val int)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@dest1
+POSTHOOK: query: -- SORT_QUERY_RESULTS
+
+create table dest1(key int, val int)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@dest1
+PREHOOK: query: explain
+insert overwrite table dest1
+select key, count(1) from src group by key
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+insert overwrite table dest1
+select key, count(1) from src group by key
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+ Stage-1 is a root stage
+ Stage-2 depends on stages: Stage-1
+ Stage-0 depends on stages: Stage-2
+ Stage-3 depends on stages: Stage-0
+
+STAGE PLANS:
+ Stage: Stage-1
+ Tez
+ Edges:
+ Reducer 2 <- Map 1 (SIMPLE_EDGE)
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string)
+ outputColumnNames: _col0
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Group By Operator
+ aggregations: count(1)
+ keys: _col0 (type: string)
+ mode: hash
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: string)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col1 (type: bigint)
+ Execution mode: llap
+ Reducer 2
+ Execution mode: uber
+ Reduce Operator Tree:
+ Group By Operator
+ aggregations: count(VALUE._col0)
+ keys: KEY._col0 (type: string)
+ mode: mergepartial
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.dest1
+
+ Stage: Stage-2
+ Dependency Collection
+
+ Stage: Stage-0
+ Move Operator
+ tables:
+ replace: true
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.dest1
+
+ Stage: Stage-3
+ Stats-Aggr Operator
+
+PREHOOK: query: insert overwrite table dest1
+select key, count(1) from src group by key
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@dest1
+POSTHOOK: query: insert overwrite table dest1
+select key, count(1) from src group by key
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@dest1
+POSTHOOK: Lineage: dest1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: dest1.val EXPRESSION [(src)src.null, ]
+PREHOOK: query: select * from dest1
+PREHOOK: type: QUERY
+PREHOOK: Input: default@dest1
+#### A masked pattern was here ####
+POSTHOOK: query: select * from dest1
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@dest1
+#### A masked pattern was here ####
+0 3
+10 1
+100 2
+103 2
+104 2
+105 1
+11 1
+111 1
+113 2
+114 1
+116 1
+118 2
+119 3
+12 2
+120 2
+125 2
+126 1
+128 3
+129 2
+131 1
+133 1
+134 2
+136 1
+137 2
+138 4
+143 1
+145 1
+146 2
+149 2
+15 2
+150 1
+152 2
+153 1
+155 1
+156 1
+157 1
+158 1
+160 1
+162 1
+163 1
+164 2
+165 2
+166 1
+167 3
+168 1
+169 4
+17 1
+170 1
+172 2
+174 2
+175 2
+176 2
+177 1
+178 1
+179 2
+18 2
+180 1
+181 1
+183 1
+186 1
+187 3
+189 1
+19 1
+190 1
+191 2
+192 1
+193 3
+194 1
+195 2
+196 1
+197 2
+199 3
+2 1
+20 1
+200 2
+201 1
+202 1
+203 2
+205 2
+207 2
+208 3
+209 2
+213 2
+214 1
+216 2
+217 2
+218 1
+219 2
+221 2
+222 1
+223 2
+224 2
+226 1
+228 1
+229 2
+230 5
+233 2
+235 1
+237 2
+238 2
+239 2
+24 2
+241 1
+242 2
+244 1
+247 1
+248 1
+249 1
+252 1
+255 2
+256 2
+257 1
+258 1
+26 2
+260 1
+262 1
+263 1
+265 2
+266 1
+27 1
+272 2
+273 3
+274 1
+275 1
+277 4
+278 2
+28 1
+280 2
+281 2
+282 2
+283 1
+284 1
+285 1
+286 1
+287 1
+288 2
+289 1
+291 1
+292 1
+296 1
+298 3
+30 1
+302 1
+305 1
+306 1
+307 2
+308 1
+309 2
+310 1
+311 3
+315 1
+316 3
+317 2
+318 3
+321 2
+322 2
+323 1
+325 2
+327 3
+33 1
+331 2
+332 1
+333 2
+335 1
+336 1
+338 1
+339 1
+34 1
+341 1
+342 2
+344 2
+345 1
+348 5
+35 3
+351 1
+353 2
+356 1
+360 1
+362 1
+364 1
+365 1
+366 1
+367 2
+368 1
+369 3
+37 2
+373 1
+374 1
+375 1
+377 1
+378 1
+379 1
+382 2
+384 3
+386 1
+389 1
+392 1
+393 1
+394 1
+395 2
+396 3
+397 2
+399 2
+4 1
+400 1
+401 5
+402 1
+403 3
+404 2
+406 4
+407 1
+409 3
+41 1
+411 1
+413 2
+414 2
+417 3
+418 1
+419 1
+42 2
+421 1
+424 2
+427 1
+429 2
+43 1
+430 3
+431 3
+432 1
+435 1
+436 1
+437 1
+438 3
+439 2
+44 1
+443 1
+444 1
+446 1
+448 1
+449 1
+452 1
+453 1
+454 3
+455 1
+457 1
+458 2
+459 2
+460 1
+462 2
+463 2
+466 3
+467 1
+468 4
+469 5
+47 1
+470 1
+472 1
+475 1
+477 1
+478 2
+479 1
+480 3
+481 1
+482 1
+483 1
+484 1
+485 1
+487 1
+489 4
+490 1
+491 1
+492 2
+493 1
+494 1
+495 1
+496 1
+497 1
+498 3
+5 3
+51 2
+53 1
+54 1
+57 1
+58 2
+64 1
+65 1
+66 1
+67 2
+69 1
+70 3
+72 2
+74 1
+76 2
+77 1
+78 1
+8 1
+80 1
+82 1
+83 2
+84 2
+85 1
+86 1
+87 1
+9 1
+90 3
+92 1
+95 2
+96 1
+97 2
+98 2
+PREHOOK: query: drop table dest1
+PREHOOK: type: DROPTABLE
+PREHOOK: Input: default@dest1
+PREHOOK: Output: default@dest1
+POSTHOOK: query: drop table dest1
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Input: default@dest1
+POSTHOOK: Output: default@dest1
+PREHOOK: query: create table test_src(key string, value string) partitioned by (ds string)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@test_src
+POSTHOOK: query: create table test_src(key string, value string) partitioned by (ds string)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@test_src
+PREHOOK: query: create table dest1(key string)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@dest1
+POSTHOOK: query: create table dest1(key string)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@dest1
+PREHOOK: query: insert overwrite table test_src partition(ds='101') select * from src
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@test_src@ds=101
+POSTHOOK: query: insert overwrite table test_src partition(ds='101') select * from src
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@test_src@ds=101
+POSTHOOK: Lineage: test_src PARTITION(ds=101).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: test_src PARTITION(ds=101).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
+PREHOOK: query: insert overwrite table test_src partition(ds='102') select * from src
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@test_src@ds=102
+POSTHOOK: query: insert overwrite table test_src partition(ds='102') select * from src
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@test_src@ds=102
+POSTHOOK: Lineage: test_src PARTITION(ds=102).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: test_src PARTITION(ds=102).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
+PREHOOK: query: explain
+insert overwrite table dest1 select key from test_src
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+insert overwrite table dest1 select key from test_src
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+ Stage-1 is a root stage
+ Stage-2 depends on stages: Stage-1
+ Stage-0 depends on stages: Stage-2
+ Stage-3 depends on stages: Stage-0
+
+STAGE PLANS:
+ Stage: Stage-1
+ Tez
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: test_src
+ Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string)
+ outputColumnNames: _col0
+ Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.dest1
+ Execution mode: llap
+
+ Stage: Stage-2
+ Dependency Collection
+
+ Stage: Stage-0
+ Move Operator
+ tables:
+ replace: true
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.dest1
+
+ Stage: Stage-3
+ Stats-Aggr Operator
+
+PREHOOK: query: insert overwrite table dest1 select key from test_src
+PREHOOK: type: QUERY
+PREHOOK: Input: default@test_src
+PREHOOK: Input: default@test_src@ds=101
+PREHOOK: Input: default@test_src@ds=102
+PREHOOK: Output: default@dest1
+POSTHOOK: query: insert overwrite table dest1 select key from test_src
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@test_src
+POSTHOOK: Input: default@test_src@ds=101
+POSTHOOK: Input: default@test_src@ds=102
+POSTHOOK: Output: default@dest1
+POSTHOOK: Lineage: dest1.key SIMPLE [(test_src)test_src.FieldSchema(name:key, type:string, comment:null), ]
+PREHOOK: query: explain
+insert overwrite table dest1 select key from test_src
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+insert overwrite table dest1 select key from test_src
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+ Stage-1 is a root stage
+ Stage-2 depends on stages: Stage-1
+ Stage-0 depends on stages: Stage-2
+ Stage-3 depends on stages: Stage-0
+
+STAGE PLANS:
+ Stage: Stage-1
+ Tez
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: test_src
+ Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string)
+ outputColumnNames: _col0
+ Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.dest1
+ Execution mode: llap
+
+ Stage: Stage-2
+ Dependency Collection
+
+ Stage: Stage-0
+ Move Operator
+ tables:
+ replace: true
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.dest1
+
+ Stage: Stage-3
+ Stats-Aggr Operator
+
+PREHOOK: query: insert overwrite table dest1 select key from test_src
+PREHOOK: type: QUERY
+PREHOOK: Input: default@test_src
+PREHOOK: Input: default@test_src@ds=101
+PREHOOK: Input: default@test_src@ds=102
+PREHOOK: Output: default@dest1
+POSTHOOK: query: insert overwrite table dest1 select key from test_src
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@test_src
+POSTHOOK: Input: default@test_src@ds=101
+POSTHOOK: Input: default@test_src@ds=102
+POSTHOOK: Output: default@dest1
+POSTHOOK: Lineage: dest1.key SIMPLE [(test_src)test_src.FieldSchema(name:key, type:string, comment:null), ]