You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by br...@apache.org on 2014/10/06 05:44:26 UTC
svn commit: r1629562 [30/38] - in /hive/branches/spark: ./ accumulo-handler/
beeline/ beeline/src/java/org/apache/hive/beeline/ bin/ext/ common/
common/src/java/org/apache/hadoop/hive/conf/
common/src/test/org/apache/hadoop/hive/common/type/ contrib/sr...
Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/dynamic_partition_pruning_2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/dynamic_partition_pruning_2.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/dynamic_partition_pruning_2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/dynamic_partition_pruning_2.q.out Mon Oct 6 03:44:13 2014
@@ -67,34 +67,6 @@ POSTHOOK: query: load data local inpath
POSTHOOK: type: LOAD
#### A masked pattern was here ####
POSTHOOK: Output: default@agg_01@dim_shops_id=3
-PREHOOK: query: analyze table dim_shops compute statistics
-PREHOOK: type: QUERY
-PREHOOK: Input: default@dim_shops
-PREHOOK: Output: default@dim_shops
-POSTHOOK: query: analyze table dim_shops compute statistics
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@dim_shops
-POSTHOOK: Output: default@dim_shops
-PREHOOK: query: analyze table agg_01 partition (dim_shops_id) compute statistics
-PREHOOK: type: QUERY
-PREHOOK: Input: default@agg_01
-PREHOOK: Input: default@agg_01@dim_shops_id=1
-PREHOOK: Input: default@agg_01@dim_shops_id=2
-PREHOOK: Input: default@agg_01@dim_shops_id=3
-PREHOOK: Output: default@agg_01
-PREHOOK: Output: default@agg_01@dim_shops_id=1
-PREHOOK: Output: default@agg_01@dim_shops_id=2
-PREHOOK: Output: default@agg_01@dim_shops_id=3
-POSTHOOK: query: analyze table agg_01 partition (dim_shops_id) compute statistics
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@agg_01
-POSTHOOK: Input: default@agg_01@dim_shops_id=1
-POSTHOOK: Input: default@agg_01@dim_shops_id=2
-POSTHOOK: Input: default@agg_01@dim_shops_id=3
-POSTHOOK: Output: default@agg_01
-POSTHOOK: Output: default@agg_01@dim_shops_id=1
-POSTHOOK: Output: default@agg_01@dim_shops_id=2
-POSTHOOK: Output: default@agg_01@dim_shops_id=3
PREHOOK: query: select * from dim_shops
PREHOOK: type: QUERY
PREHOOK: Input: default@dim_shops
@@ -165,556 +137,29 @@ STAGE PLANS:
TableScan
alias: d1
filterExpr: (id is not null and (label) IN ('foo', 'bar')) (type: boolean)
- Statistics: Num rows: 3 Data size: 15 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 18 Basic stats: PARTIAL Column stats: NONE
Filter Operator
predicate: (id is not null and (label) IN ('foo', 'bar')) (type: boolean)
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Reduce Output Operator
key expressions: id (type: int)
sort order: +
Map-reduce partition columns: id (type: int)
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
- value expressions: label (type: string)
- Select Operator
- expressions: id (type: int)
- outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- keys: _col0 (type: int)
- mode: hash
- outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
- Dynamic Partitioning Event Operator
- Target Input: agg
- Partition key expr: dim_shops_id
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
- Target column: dim_shops_id
- Target Vertex: Map 2
- Map 2
- Map Operator Tree:
- TableScan
- alias: agg
- filterExpr: dim_shops_id is not null (type: boolean)
- Statistics: Num rows: 9 Data size: 27 Basic stats: COMPLETE Column stats: NONE
- Map Join Operator
- condition map:
- Inner Join 0 to 1
- condition expressions:
- 0 {amount} {dim_shops_id}
- 1 {id} {label}
- keys:
- 0 dim_shops_id (type: int)
- 1 id (type: int)
- outputColumnNames: _col0, _col1, _col5, _col6
- input vertices:
- 1 Map 1
- Statistics: Num rows: 9 Data size: 29 Basic stats: COMPLETE Column stats: NONE
- Filter Operator
- predicate: ((_col1 = _col5) and (_col6) IN ('foo', 'bar')) (type: boolean)
- Statistics: Num rows: 2 Data size: 6 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col6 (type: string), _col0 (type: decimal(10,0))
- outputColumnNames: _col6, _col0
- Statistics: Num rows: 2 Data size: 6 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- aggregations: count(), sum(_col0)
- keys: _col6 (type: string)
- mode: hash
- outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 2 Data size: 6 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: string)
- sort order: +
- Map-reduce partition columns: _col0 (type: string)
- Statistics: Num rows: 2 Data size: 6 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col1 (type: bigint), _col2 (type: decimal(20,0))
- Reducer 3
- Reduce Operator Tree:
- Group By Operator
- aggregations: count(VALUE._col0), sum(VALUE._col1)
- keys: KEY._col0 (type: string)
- mode: mergepartial
- outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: bigint), _col2 (type: decimal(20,0))
- outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: string)
- sort order: +
- Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col1 (type: bigint), _col2 (type: decimal(20,0))
- Reducer 4
- Reduce Operator Tree:
- Select Operator
- expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: bigint), VALUE._col1 (type: decimal(20,0))
- outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-
- Stage: Stage-0
- Fetch Operator
- limit: -1
- Processor Tree:
- ListSink
-
-PREHOOK: query: SELECT d1.label, count(*), sum(agg.amount)
-FROM agg_01 agg,
-dim_shops d1
-WHERE agg.dim_shops_id = d1.id
-and
-d1.label in ('foo', 'bar')
-GROUP BY d1.label
-ORDER BY d1.label
-PREHOOK: type: QUERY
-PREHOOK: Input: default@agg_01
-PREHOOK: Input: default@agg_01@dim_shops_id=1
-PREHOOK: Input: default@agg_01@dim_shops_id=2
-PREHOOK: Input: default@agg_01@dim_shops_id=3
-PREHOOK: Input: default@dim_shops
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT d1.label, count(*), sum(agg.amount)
-FROM agg_01 agg,
-dim_shops d1
-WHERE agg.dim_shops_id = d1.id
-and
-d1.label in ('foo', 'bar')
-GROUP BY d1.label
-ORDER BY d1.label
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@agg_01
-POSTHOOK: Input: default@agg_01@dim_shops_id=1
-POSTHOOK: Input: default@agg_01@dim_shops_id=2
-POSTHOOK: Input: default@agg_01@dim_shops_id=3
-POSTHOOK: Input: default@dim_shops
-#### A masked pattern was here ####
-bar 3 15
-foo 3 6
-PREHOOK: query: EXPLAIN SELECT d1.label, count(*), sum(agg.amount)
-FROM agg_01 agg,
-dim_shops d1
-WHERE agg.dim_shops_id = d1.id
-and
-d1.label in ('foo', 'bar')
-GROUP BY d1.label
-ORDER BY d1.label
-PREHOOK: type: QUERY
-POSTHOOK: query: EXPLAIN SELECT d1.label, count(*), sum(agg.amount)
-FROM agg_01 agg,
-dim_shops d1
-WHERE agg.dim_shops_id = d1.id
-and
-d1.label in ('foo', 'bar')
-GROUP BY d1.label
-ORDER BY d1.label
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
- Stage: Stage-1
- Tez
- Edges:
- Map 2 <- Map 1 (BROADCAST_EDGE)
- Reducer 3 <- Map 2 (SIMPLE_EDGE)
- Reducer 4 <- Reducer 3 (SIMPLE_EDGE)
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: d1
- filterExpr: (id is not null and (label) IN ('foo', 'bar')) (type: boolean)
- Statistics: Num rows: 3 Data size: 15 Basic stats: COMPLETE Column stats: NONE
- Filter Operator
- predicate: (id is not null and (label) IN ('foo', 'bar')) (type: boolean)
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: id (type: int)
- sort order: +
- Map-reduce partition columns: id (type: int)
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
- value expressions: label (type: string)
- Map 2
- Map Operator Tree:
- TableScan
- alias: agg
- filterExpr: dim_shops_id is not null (type: boolean)
- Statistics: Num rows: 9 Data size: 27 Basic stats: COMPLETE Column stats: NONE
- Map Join Operator
- condition map:
- Inner Join 0 to 1
- condition expressions:
- 0 {amount} {dim_shops_id}
- 1 {id} {label}
- keys:
- 0 dim_shops_id (type: int)
- 1 id (type: int)
- outputColumnNames: _col0, _col1, _col5, _col6
- input vertices:
- 1 Map 1
- Statistics: Num rows: 9 Data size: 29 Basic stats: COMPLETE Column stats: NONE
- Filter Operator
- predicate: ((_col1 = _col5) and (_col6) IN ('foo', 'bar')) (type: boolean)
- Statistics: Num rows: 2 Data size: 6 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col6 (type: string), _col0 (type: decimal(10,0))
- outputColumnNames: _col6, _col0
- Statistics: Num rows: 2 Data size: 6 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- aggregations: count(), sum(_col0)
- keys: _col6 (type: string)
- mode: hash
- outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 2 Data size: 6 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: string)
- sort order: +
- Map-reduce partition columns: _col0 (type: string)
- Statistics: Num rows: 2 Data size: 6 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col1 (type: bigint), _col2 (type: decimal(20,0))
- Reducer 3
- Reduce Operator Tree:
- Group By Operator
- aggregations: count(VALUE._col0), sum(VALUE._col1)
- keys: KEY._col0 (type: string)
- mode: mergepartial
- outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: bigint), _col2 (type: decimal(20,0))
- outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: string)
- sort order: +
- Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col1 (type: bigint), _col2 (type: decimal(20,0))
- Reducer 4
- Reduce Operator Tree:
- Select Operator
- expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: bigint), VALUE._col1 (type: decimal(20,0))
- outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-
- Stage: Stage-0
- Fetch Operator
- limit: -1
- Processor Tree:
- ListSink
-
-PREHOOK: query: SELECT d1.label, count(*), sum(agg.amount)
-FROM agg_01 agg,
-dim_shops d1
-WHERE agg.dim_shops_id = d1.id
-and
-d1.label in ('foo', 'bar')
-GROUP BY d1.label
-ORDER BY d1.label
-PREHOOK: type: QUERY
-PREHOOK: Input: default@agg_01
-PREHOOK: Input: default@agg_01@dim_shops_id=1
-PREHOOK: Input: default@agg_01@dim_shops_id=2
-PREHOOK: Input: default@agg_01@dim_shops_id=3
-PREHOOK: Input: default@dim_shops
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT d1.label, count(*), sum(agg.amount)
-FROM agg_01 agg,
-dim_shops d1
-WHERE agg.dim_shops_id = d1.id
-and
-d1.label in ('foo', 'bar')
-GROUP BY d1.label
-ORDER BY d1.label
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@agg_01
-POSTHOOK: Input: default@agg_01@dim_shops_id=1
-POSTHOOK: Input: default@agg_01@dim_shops_id=2
-POSTHOOK: Input: default@agg_01@dim_shops_id=3
-POSTHOOK: Input: default@dim_shops
-#### A masked pattern was here ####
-bar 3 15
-foo 3 6
-PREHOOK: query: EXPLAIN SELECT d1.label
-FROM agg_01 agg,
-dim_shops d1
-WHERE agg.dim_shops_id = d1.id
-PREHOOK: type: QUERY
-POSTHOOK: query: EXPLAIN SELECT d1.label
-FROM agg_01 agg,
-dim_shops d1
-WHERE agg.dim_shops_id = d1.id
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
- Stage: Stage-1
- Tez
- Edges:
- Map 2 <- Map 1 (BROADCAST_EDGE)
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: d1
- filterExpr: id is not null (type: boolean)
- Statistics: Num rows: 3 Data size: 15 Basic stats: COMPLETE Column stats: NONE
- Filter Operator
- predicate: id is not null (type: boolean)
- Statistics: Num rows: 2 Data size: 10 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: id (type: int)
- sort order: +
- Map-reduce partition columns: id (type: int)
- Statistics: Num rows: 2 Data size: 10 Basic stats: COMPLETE Column stats: NONE
- value expressions: label (type: string)
- Map 2
- Map Operator Tree:
- TableScan
- alias: agg
- filterExpr: dim_shops_id is not null (type: boolean)
- Statistics: Num rows: 9 Data size: 27 Basic stats: COMPLETE Column stats: NONE
- Map Join Operator
- condition map:
- Inner Join 0 to 1
- condition expressions:
- 0 {dim_shops_id}
- 1 {id} {label}
- keys:
- 0 dim_shops_id (type: int)
- 1 id (type: int)
- outputColumnNames: _col1, _col5, _col6
- input vertices:
- 1 Map 1
- Statistics: Num rows: 9 Data size: 29 Basic stats: COMPLETE Column stats: NONE
- Filter Operator
- predicate: (_col1 = _col5) (type: boolean)
- Statistics: Num rows: 4 Data size: 12 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col6 (type: string)
- outputColumnNames: _col0
- Statistics: Num rows: 4 Data size: 12 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 4 Data size: 12 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-
- Stage: Stage-0
- Fetch Operator
- limit: -1
- Processor Tree:
- ListSink
-
-PREHOOK: query: SELECT d1.label
-FROM agg_01 agg,
-dim_shops d1
-WHERE agg.dim_shops_id = d1.id
-PREHOOK: type: QUERY
-PREHOOK: Input: default@agg_01
-PREHOOK: Input: default@agg_01@dim_shops_id=1
-PREHOOK: Input: default@agg_01@dim_shops_id=2
-PREHOOK: Input: default@agg_01@dim_shops_id=3
-PREHOOK: Input: default@dim_shops
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT d1.label
-FROM agg_01 agg,
-dim_shops d1
-WHERE agg.dim_shops_id = d1.id
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@agg_01
-POSTHOOK: Input: default@agg_01@dim_shops_id=1
-POSTHOOK: Input: default@agg_01@dim_shops_id=2
-POSTHOOK: Input: default@agg_01@dim_shops_id=3
-POSTHOOK: Input: default@dim_shops
-#### A masked pattern was here ####
-foo
-foo
-foo
-bar
-bar
-bar
-baz
-baz
-baz
-PREHOOK: query: EXPLAIN SELECT agg.amount
-FROM agg_01 agg,
-dim_shops d1
-WHERE agg.dim_shops_id = d1.id
-and agg.dim_shops_id = 1
-PREHOOK: type: QUERY
-POSTHOOK: query: EXPLAIN SELECT agg.amount
-FROM agg_01 agg,
-dim_shops d1
-WHERE agg.dim_shops_id = d1.id
-and agg.dim_shops_id = 1
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
- Stage: Stage-1
- Tez
- Edges:
- Map 2 <- Map 1 (BROADCAST_EDGE)
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: d1
- filterExpr: (id is not null and (id = 1)) (type: boolean)
- Statistics: Num rows: 3 Data size: 15 Basic stats: COMPLETE Column stats: NONE
- Filter Operator
- predicate: (id is not null and (id = 1)) (type: boolean)
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: 1 (type: int)
- sort order: +
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
- Map 2
- Map Operator Tree:
- TableScan
- alias: agg
- filterExpr: (dim_shops_id is not null and (dim_shops_id = 1)) (type: boolean)
- Statistics: Num rows: 3 Data size: 9 Basic stats: COMPLETE Column stats: NONE
- Map Join Operator
- condition map:
- Inner Join 0 to 1
- condition expressions:
- 0 {amount} {dim_shops_id}
- 1
- keys:
- 0 dim_shops_id (type: int)
- 1 1 (type: int)
- outputColumnNames: _col0, _col1
- input vertices:
- 1 Map 1
- Statistics: Num rows: 3 Data size: 9 Basic stats: COMPLETE Column stats: NONE
- Filter Operator
- predicate: ((_col1 = 1) and (_col1 = 1)) (type: boolean)
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
- Select Operator
- expressions: _col0 (type: decimal(10,0))
- outputColumnNames: _col0
- Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-
- Stage: Stage-0
- Fetch Operator
- limit: -1
- Processor Tree:
- ListSink
-
-PREHOOK: query: SELECT agg.amount
-FROM agg_01 agg,
-dim_shops d1
-WHERE agg.dim_shops_id = d1.id
-and agg.dim_shops_id = 1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@agg_01
-PREHOOK: Input: default@agg_01@dim_shops_id=1
-PREHOOK: Input: default@dim_shops
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT agg.amount
-FROM agg_01 agg,
-dim_shops d1
-WHERE agg.dim_shops_id = d1.id
-and agg.dim_shops_id = 1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@agg_01
-POSTHOOK: Input: default@agg_01@dim_shops_id=1
-POSTHOOK: Input: default@dim_shops
-#### A masked pattern was here ####
-1
-2
-3
-PREHOOK: query: EXPLAIN SELECT d1.label, count(*), sum(agg.amount)
-FROM agg_01 agg,
-dim_shops d1
-WHERE agg.dim_shops_id = d1.id
-and
-d1.label in ('foo', 'bar')
-GROUP BY d1.label
-ORDER BY d1.label
-PREHOOK: type: QUERY
-POSTHOOK: query: EXPLAIN SELECT d1.label, count(*), sum(agg.amount)
-FROM agg_01 agg,
-dim_shops d1
-WHERE agg.dim_shops_id = d1.id
-and
-d1.label in ('foo', 'bar')
-GROUP BY d1.label
-ORDER BY d1.label
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
- Stage: Stage-1
- Tez
- Edges:
- Map 2 <- Map 1 (BROADCAST_EDGE)
- Reducer 3 <- Map 2 (SIMPLE_EDGE)
- Reducer 4 <- Reducer 3 (SIMPLE_EDGE)
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: d1
- filterExpr: (id is not null and (label) IN ('foo', 'bar')) (type: boolean)
- Statistics: Num rows: 3 Data size: 15 Basic stats: COMPLETE Column stats: NONE
- Filter Operator
- predicate: (id is not null and (label) IN ('foo', 'bar')) (type: boolean)
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: id (type: int)
- sort order: +
- Map-reduce partition columns: id (type: int)
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
value expressions: label (type: string)
Select Operator
expressions: id (type: int)
outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Group By Operator
keys: _col0 (type: int)
mode: hash
outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Dynamic Partitioning Event Operator
Target Input: agg
Partition key expr: dim_shops_id
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Target column: dim_shops_id
Target Vertex: Map 2
Map 2
@@ -722,7 +167,7 @@ STAGE PLANS:
TableScan
alias: agg
filterExpr: dim_shops_id is not null (type: boolean)
- Statistics: Num rows: 9 Data size: 27 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 36 Basic stats: PARTIAL Column stats: NONE
Map Join Operator
condition map:
Inner Join 0 to 1
@@ -733,27 +178,25 @@ STAGE PLANS:
0 dim_shops_id (type: int)
1 id (type: int)
outputColumnNames: _col0, _col1, _col5, _col6
- input vertices:
- 1 Map 1
- Statistics: Num rows: 9 Data size: 29 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 39 Basic stats: PARTIAL Column stats: NONE
Filter Operator
predicate: ((_col1 = _col5) and (_col6) IN ('foo', 'bar')) (type: boolean)
- Statistics: Num rows: 2 Data size: 6 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: _col6 (type: string), _col0 (type: decimal(10,0))
outputColumnNames: _col6, _col0
- Statistics: Num rows: 2 Data size: 6 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Group By Operator
aggregations: count(), sum(_col0)
keys: _col6 (type: string)
mode: hash
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 2 Data size: 6 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string)
sort order: +
Map-reduce partition columns: _col0 (type: string)
- Statistics: Num rows: 2 Data size: 6 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
value expressions: _col1 (type: bigint), _col2 (type: decimal(20,0))
Reducer 3
Reduce Operator Tree:
@@ -762,25 +205,25 @@ STAGE PLANS:
keys: KEY._col0 (type: string)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: _col0 (type: string), _col1 (type: bigint), _col2 (type: decimal(20,0))
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string)
sort order: +
- Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
value expressions: _col1 (type: bigint), _col2 (type: decimal(20,0))
Reducer 4
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: bigint), VALUE._col1 (type: decimal(20,0))
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -851,28 +294,28 @@ STAGE PLANS:
TableScan
alias: dim_shops
filterExpr: (id is not null and (label = 'bar')) (type: boolean)
- Statistics: Num rows: 3 Data size: 15 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 2 Data size: 18 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: (id is not null and (label = 'bar')) (type: boolean)
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Reduce Output Operator
key expressions: id (type: int)
sort order: +
Map-reduce partition columns: id (type: int)
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: id (type: int)
outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Group By Operator
keys: _col0 (type: int)
mode: hash
outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Dynamic Partitioning Event Operator
Target Input: agg_01
Partition key expr: dim_shops_id
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Target column: dim_shops_id
Target Vertex: Map 3
Map 2
@@ -880,28 +323,28 @@ STAGE PLANS:
TableScan
alias: dim_shops
filterExpr: (id is not null and (label = 'foo')) (type: boolean)
- Statistics: Num rows: 3 Data size: 15 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 2 Data size: 18 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: (id is not null and (label = 'foo')) (type: boolean)
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Reduce Output Operator
key expressions: id (type: int)
sort order: +
Map-reduce partition columns: id (type: int)
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: id (type: int)
outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Group By Operator
keys: _col0 (type: int)
mode: hash
outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Dynamic Partitioning Event Operator
Target Input: agg_01
Partition key expr: dim_shops_id
- Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Target column: dim_shops_id
Target Vertex: Map 5
Map 3
@@ -919,8 +362,6 @@ STAGE PLANS:
0 dim_shops_id (type: int)
1 id (type: int)
outputColumnNames: _col0, _col1, _col5
- input vertices:
- 1 Map 1
Filter Operator
predicate: (_col1 = _col5) (type: boolean)
Select Operator
@@ -950,8 +391,6 @@ STAGE PLANS:
0 dim_shops_id (type: int)
1 id (type: int)
outputColumnNames: _col0, _col1, _col5
- input vertices:
- 1 Map 2
Filter Operator
predicate: (_col1 = _col5) (type: boolean)
Select Operator
Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out Mon Oct 6 03:44:13 2014
@@ -214,7 +214,6 @@ STAGE PLANS:
output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.over1k_part_orc
- Execution mode: vectorized
Stage: Stage-2
Dependency Collection
@@ -301,7 +300,6 @@ STAGE PLANS:
output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.over1k_part_limit_orc
- Execution mode: vectorized
Stage: Stage-2
Dependency Collection
@@ -370,7 +368,6 @@ STAGE PLANS:
output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.over1k_part_buck_orc
- Execution mode: vectorized
Stage: Stage-2
Dependency Collection
@@ -438,7 +435,6 @@ STAGE PLANS:
output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.over1k_part_buck_sort_orc
- Execution mode: vectorized
Stage: Stage-2
Dependency Collection
@@ -589,7 +585,6 @@ STAGE PLANS:
output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.over1k_part_orc
- Execution mode: vectorized
Stage: Stage-2
Dependency Collection
@@ -676,7 +671,6 @@ STAGE PLANS:
output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.over1k_part_limit_orc
- Execution mode: vectorized
Stage: Stage-2
Dependency Collection
@@ -745,7 +739,6 @@ STAGE PLANS:
output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.over1k_part_buck_orc
- Execution mode: vectorized
Stage: Stage-2
Dependency Collection
@@ -813,7 +806,6 @@ STAGE PLANS:
output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.over1k_part_buck_sort_orc
- Execution mode: vectorized
Stage: Stage-2
Dependency Collection
@@ -976,7 +968,7 @@ Partition Parameters:
numFiles 2
numRows 3
rawDataSize 60
- totalSize 1050
+ totalSize 1034
#### A masked pattern was here ####
# Storage Information
@@ -1064,7 +1056,7 @@ Partition Parameters:
numFiles 2
numRows 3
rawDataSize 60
- totalSize 1050
+ totalSize 1040
#### A masked pattern was here ####
# Storage Information
@@ -1107,7 +1099,7 @@ Partition Parameters:
numFiles 8
numRows 16
rawDataSize 320
- totalSize 4340
+ totalSize 4332
#### A masked pattern was here ####
# Storage Information
@@ -1193,7 +1185,7 @@ Partition Parameters:
numFiles 8
numRows 16
rawDataSize 320
- totalSize 4326
+ totalSize 4318
#### A masked pattern was here ####
# Storage Information
@@ -1370,7 +1362,6 @@ STAGE PLANS:
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.over1k_part2_orc
- Execution mode: vectorized
Stage: Stage-2
Dependency Collection
@@ -1452,270 +1443,6 @@ STAGE PLANS:
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.over1k_part2_orc
- Execution mode: vectorized
-
- Stage: Stage-2
- Dependency Collection
-
- Stage: Stage-0
- Move Operator
- tables:
- partition:
- ds foo
- t
- replace: true
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.over1k_part2_orc
-
- Stage: Stage-3
- Stats-Aggr Operator
-
-PREHOOK: query: explain insert overwrite table over1k_part2_orc partition(ds="foo",t) select si,i,b,f,t from (select * from over1k_orc order by i limit 10) tmp where t is null or t=27
-PREHOOK: type: QUERY
-POSTHOOK: query: explain insert overwrite table over1k_part2_orc partition(ds="foo",t) select si,i,b,f,t from (select * from over1k_orc order by i limit 10) tmp where t is null or t=27
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-2 depends on stages: Stage-1
- Stage-0 depends on stages: Stage-2
- Stage-3 depends on stages: Stage-0
-
-STAGE PLANS:
- Stage: Stage-1
- Tez
- Edges:
- Reducer 2 <- Map 1 (SIMPLE_EDGE)
- Reducer 3 <- Reducer 2 (SIMPLE_EDGE)
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: over1k_orc
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: t (type: tinyint), si (type: smallint), i (type: int), b (type: bigint), f (type: float)
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col2 (type: int)
- sort order: +
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col0 (type: tinyint), _col1 (type: smallint), _col3 (type: bigint), _col4 (type: float)
- Execution mode: vectorized
- Reducer 2
- Reduce Operator Tree:
- Select Operator
- expressions: VALUE._col0 (type: tinyint), VALUE._col1 (type: smallint), KEY.reducesinkkey0 (type: int), VALUE._col2 (type: bigint), VALUE._col3 (type: float)
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
- Limit
- Number of rows: 10
- Statistics: Num rows: 10 Data size: 2960 Basic stats: COMPLETE Column stats: NONE
- Filter Operator
- predicate: (_col0 is null or (_col0 = 27)) (type: boolean)
- Statistics: Num rows: 10 Data size: 2960 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col1 (type: smallint), _col2 (type: int), _col3 (type: bigint), _col4 (type: float), _col0 (type: tinyint)
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 10 Data size: 2960 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col4 (type: tinyint)
- sort order: +
- Map-reduce partition columns: _col4 (type: tinyint)
- Statistics: Num rows: 10 Data size: 2960 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint)
- Execution mode: vectorized
- Reducer 3
- Reduce Operator Tree:
- Extract
- Statistics: Num rows: 10 Data size: 2960 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 10 Data size: 2960 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.over1k_part2_orc
- Execution mode: vectorized
-
- Stage: Stage-2
- Dependency Collection
-
- Stage: Stage-0
- Move Operator
- tables:
- partition:
- ds foo
- t
- replace: true
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.over1k_part2_orc
-
- Stage: Stage-3
- Stats-Aggr Operator
-
-PREHOOK: query: explain insert overwrite table over1k_part2_orc partition(ds="foo",t) select si,i,b,f,t from over1k_orc where t is null or t=27 group by si,i,b,f,t
-PREHOOK: type: QUERY
-POSTHOOK: query: explain insert overwrite table over1k_part2_orc partition(ds="foo",t) select si,i,b,f,t from over1k_orc where t is null or t=27 group by si,i,b,f,t
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-2 depends on stages: Stage-1
- Stage-0 depends on stages: Stage-2
- Stage-3 depends on stages: Stage-0
-
-STAGE PLANS:
- Stage: Stage-1
- Tez
- Edges:
- Reducer 2 <- Map 1 (SIMPLE_EDGE)
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: over1k_orc
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
- Filter Operator
- predicate: (t is null or (t = 27)) (type: boolean)
- Statistics: Num rows: 1048 Data size: 310873 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: si (type: smallint), i (type: int), b (type: bigint), f (type: float), t (type: tinyint)
- outputColumnNames: si, i, b, f, t
- Statistics: Num rows: 1048 Data size: 310873 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- keys: si (type: smallint), i (type: int), b (type: bigint), f (type: float), t (type: tinyint)
- mode: hash
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 1048 Data size: 310873 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint)
- sort order: +++++
- Map-reduce partition columns: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint)
- Statistics: Num rows: 1048 Data size: 310873 Basic stats: COMPLETE Column stats: NONE
- Execution mode: vectorized
- Reducer 2
- Reduce Operator Tree:
- Group By Operator
- keys: KEY._col0 (type: smallint), KEY._col1 (type: int), KEY._col2 (type: bigint), KEY._col3 (type: float), KEY._col4 (type: tinyint)
- mode: mergepartial
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 524 Data size: 155436 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint)
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 524 Data size: 155436 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 524 Data size: 155436 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.over1k_part2_orc
- Execution mode: vectorized
-
- Stage: Stage-2
- Dependency Collection
-
- Stage: Stage-0
- Move Operator
- tables:
- partition:
- ds foo
- t
- replace: true
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.over1k_part2_orc
-
- Stage: Stage-3
- Stats-Aggr Operator
-
-PREHOOK: query: -- tests for HIVE-8162, only partition column 't' should be in last RS operator
-explain insert overwrite table over1k_part2_orc partition(ds="foo",t) select si,i,b,f,t from over1k_orc where t is null or t=27 group by si,i,b,f,t
-PREHOOK: type: QUERY
-POSTHOOK: query: -- tests for HIVE-8162, only partition column 't' should be in last RS operator
-explain insert overwrite table over1k_part2_orc partition(ds="foo",t) select si,i,b,f,t from over1k_orc where t is null or t=27 group by si,i,b,f,t
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-2 depends on stages: Stage-1
- Stage-0 depends on stages: Stage-2
- Stage-3 depends on stages: Stage-0
-
-STAGE PLANS:
- Stage: Stage-1
- Tez
- Edges:
- Reducer 2 <- Map 1 (SIMPLE_EDGE)
- Reducer 3 <- Reducer 2 (SIMPLE_EDGE)
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: over1k_orc
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
- Filter Operator
- predicate: (t is null or (t = 27)) (type: boolean)
- Statistics: Num rows: 1048 Data size: 310873 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: si (type: smallint), i (type: int), b (type: bigint), f (type: float), t (type: tinyint)
- outputColumnNames: si, i, b, f, t
- Statistics: Num rows: 1048 Data size: 310873 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- keys: si (type: smallint), i (type: int), b (type: bigint), f (type: float), t (type: tinyint)
- mode: hash
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 1048 Data size: 310873 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint)
- sort order: +++++
- Map-reduce partition columns: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint)
- Statistics: Num rows: 1048 Data size: 310873 Basic stats: COMPLETE Column stats: NONE
- Execution mode: vectorized
- Reducer 2
- Reduce Operator Tree:
- Group By Operator
- keys: KEY._col0 (type: smallint), KEY._col1 (type: int), KEY._col2 (type: bigint), KEY._col3 (type: float), KEY._col4 (type: tinyint)
- mode: mergepartial
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 524 Data size: 155436 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint)
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 524 Data size: 155436 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col4 (type: tinyint)
- sort order: +
- Map-reduce partition columns: _col4 (type: tinyint)
- Statistics: Num rows: 524 Data size: 155436 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint)
- Execution mode: vectorized
- Reducer 3
- Reduce Operator Tree:
- Extract
- Statistics: Num rows: 524 Data size: 155436 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 524 Data size: 155436 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.over1k_part2_orc
- Execution mode: vectorized
Stage: Stage-2
Dependency Collection
@@ -2110,7 +1837,6 @@ STAGE PLANS:
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.over1k_part_buck_sort2_orc
- Execution mode: vectorized
Stage: Stage-2
Dependency Collection
@@ -2178,7 +1904,6 @@ STAGE PLANS:
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.over1k_part_buck_sort2_orc
- Execution mode: vectorized
Stage: Stage-2
Dependency Collection
@@ -2370,17 +2095,17 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: over1k_part_buck_sort2_orc
- Statistics: Num rows: 19 Data size: 493 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 19 Data size: 493 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
- Statistics: Num rows: 19 Data size: 493 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 19 Data size: 493 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: count()
mode: hash
outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
sort order:
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
value expressions: _col0 (type: bigint)
Reducer 2
Reduce Operator Tree:
@@ -2388,14 +2113,14 @@ STAGE PLANS:
aggregations: count(VALUE._col0)
mode: mergepartial
outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col0 (type: bigint)
outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -2593,17 +2318,17 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: over1k_part_buck_sort2_orc
- Statistics: Num rows: 19 Data size: 493 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 19 Data size: 493 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
- Statistics: Num rows: 19 Data size: 493 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 19 Data size: 493 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: count()
mode: hash
outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
sort order:
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
value expressions: _col0 (type: bigint)
Reducer 2
Reduce Operator Tree:
@@ -2611,14 +2336,14 @@ STAGE PLANS:
aggregations: count(VALUE._col0)
mode: mergepartial
outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col0 (type: bigint)
outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/dynpart_sort_optimization.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/dynpart_sort_optimization.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/dynpart_sort_optimization.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/dynpart_sort_optimization.q.out Mon Oct 6 03:44:13 2014
@@ -1381,261 +1381,6 @@ STAGE PLANS:
Stage: Stage-3
Stats-Aggr Operator
-PREHOOK: query: explain insert overwrite table over1k_part2 partition(ds="foo",t) select si,i,b,f,t from (select * from over1k order by i limit 10) tmp where t is null or t=27
-PREHOOK: type: QUERY
-POSTHOOK: query: explain insert overwrite table over1k_part2 partition(ds="foo",t) select si,i,b,f,t from (select * from over1k order by i limit 10) tmp where t is null or t=27
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-2 depends on stages: Stage-1
- Stage-0 depends on stages: Stage-2
- Stage-3 depends on stages: Stage-0
-
-STAGE PLANS:
- Stage: Stage-1
- Tez
- Edges:
- Reducer 2 <- Map 1 (SIMPLE_EDGE)
- Reducer 3 <- Reducer 2 (SIMPLE_EDGE)
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: over1k
- Statistics: Num rows: 4443 Data size: 106636 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: t (type: tinyint), si (type: smallint), i (type: int), b (type: bigint), f (type: float)
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 4443 Data size: 106636 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col2 (type: int)
- sort order: +
- Statistics: Num rows: 4443 Data size: 106636 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col0 (type: tinyint), _col1 (type: smallint), _col3 (type: bigint), _col4 (type: float)
- Reducer 2
- Reduce Operator Tree:
- Select Operator
- expressions: VALUE._col0 (type: tinyint), VALUE._col1 (type: smallint), KEY.reducesinkkey0 (type: int), VALUE._col2 (type: bigint), VALUE._col3 (type: float)
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 4443 Data size: 106636 Basic stats: COMPLETE Column stats: NONE
- Limit
- Number of rows: 10
- Statistics: Num rows: 10 Data size: 240 Basic stats: COMPLETE Column stats: NONE
- Filter Operator
- predicate: (_col0 is null or (_col0 = 27)) (type: boolean)
- Statistics: Num rows: 10 Data size: 240 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col1 (type: smallint), _col2 (type: int), _col3 (type: bigint), _col4 (type: float), _col0 (type: tinyint)
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 10 Data size: 240 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col4 (type: tinyint)
- sort order: +
- Map-reduce partition columns: _col4 (type: tinyint)
- Statistics: Num rows: 10 Data size: 240 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint)
- Reducer 3
- Reduce Operator Tree:
- Extract
- Statistics: Num rows: 10 Data size: 240 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 10 Data size: 240 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.over1k_part2
-
- Stage: Stage-2
- Dependency Collection
-
- Stage: Stage-0
- Move Operator
- tables:
- partition:
- ds foo
- t
- replace: true
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.over1k_part2
-
- Stage: Stage-3
- Stats-Aggr Operator
-
-PREHOOK: query: explain insert overwrite table over1k_part2 partition(ds="foo",t) select si,i,b,f,t from over1k where t is null or t=27 group by si,i,b,f,t
-PREHOOK: type: QUERY
-POSTHOOK: query: explain insert overwrite table over1k_part2 partition(ds="foo",t) select si,i,b,f,t from over1k where t is null or t=27 group by si,i,b,f,t
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-2 depends on stages: Stage-1
- Stage-0 depends on stages: Stage-2
- Stage-3 depends on stages: Stage-0
-
-STAGE PLANS:
- Stage: Stage-1
- Tez
- Edges:
- Reducer 2 <- Map 1 (SIMPLE_EDGE)
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: over1k
- Statistics: Num rows: 4443 Data size: 106636 Basic stats: COMPLETE Column stats: NONE
- Filter Operator
- predicate: (t is null or (t = 27)) (type: boolean)
- Statistics: Num rows: 4442 Data size: 106611 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: si (type: smallint), i (type: int), b (type: bigint), f (type: float), t (type: tinyint)
- outputColumnNames: si, i, b, f, t
- Statistics: Num rows: 4442 Data size: 106611 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- keys: si (type: smallint), i (type: int), b (type: bigint), f (type: float), t (type: tinyint)
- mode: hash
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 4442 Data size: 106611 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint)
- sort order: +++++
- Map-reduce partition columns: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint)
- Statistics: Num rows: 4442 Data size: 106611 Basic stats: COMPLETE Column stats: NONE
- Reducer 2
- Reduce Operator Tree:
- Group By Operator
- keys: KEY._col0 (type: smallint), KEY._col1 (type: int), KEY._col2 (type: bigint), KEY._col3 (type: float), KEY._col4 (type: tinyint)
- mode: mergepartial
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 2221 Data size: 53305 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint)
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 2221 Data size: 53305 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 2221 Data size: 53305 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.over1k_part2
-
- Stage: Stage-2
- Dependency Collection
-
- Stage: Stage-0
- Move Operator
- tables:
- partition:
- ds foo
- t
- replace: true
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.over1k_part2
-
- Stage: Stage-3
- Stats-Aggr Operator
-
-PREHOOK: query: -- tests for HIVE-8162, only partition column 't' should be in last RS operator
-explain insert overwrite table over1k_part2 partition(ds="foo",t) select si,i,b,f,t from over1k where t is null or t=27 group by si,i,b,f,t
-PREHOOK: type: QUERY
-POSTHOOK: query: -- tests for HIVE-8162, only partition column 't' should be in last RS operator
-explain insert overwrite table over1k_part2 partition(ds="foo",t) select si,i,b,f,t from over1k where t is null or t=27 group by si,i,b,f,t
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-2 depends on stages: Stage-1
- Stage-0 depends on stages: Stage-2
- Stage-3 depends on stages: Stage-0
-
-STAGE PLANS:
- Stage: Stage-1
- Tez
- Edges:
- Reducer 2 <- Map 1 (SIMPLE_EDGE)
- Reducer 3 <- Reducer 2 (SIMPLE_EDGE)
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: over1k
- Statistics: Num rows: 4443 Data size: 106636 Basic stats: COMPLETE Column stats: NONE
- Filter Operator
- predicate: (t is null or (t = 27)) (type: boolean)
- Statistics: Num rows: 4442 Data size: 106611 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: si (type: smallint), i (type: int), b (type: bigint), f (type: float), t (type: tinyint)
- outputColumnNames: si, i, b, f, t
- Statistics: Num rows: 4442 Data size: 106611 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- keys: si (type: smallint), i (type: int), b (type: bigint), f (type: float), t (type: tinyint)
- mode: hash
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 4442 Data size: 106611 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint)
- sort order: +++++
- Map-reduce partition columns: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint)
- Statistics: Num rows: 4442 Data size: 106611 Basic stats: COMPLETE Column stats: NONE
- Reducer 2
- Reduce Operator Tree:
- Group By Operator
- keys: KEY._col0 (type: smallint), KEY._col1 (type: int), KEY._col2 (type: bigint), KEY._col3 (type: float), KEY._col4 (type: tinyint)
- mode: mergepartial
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 2221 Data size: 53305 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint)
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 2221 Data size: 53305 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col4 (type: tinyint)
- sort order: +
- Map-reduce partition columns: _col4 (type: tinyint)
- Statistics: Num rows: 2221 Data size: 53305 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint)
- Reducer 3
- Reduce Operator Tree:
- Extract
- Statistics: Num rows: 2221 Data size: 53305 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 2221 Data size: 53305 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.over1k_part2
-
- Stage: Stage-2
- Dependency Collection
-
- Stage: Stage-0
- Move Operator
- tables:
- partition:
- ds foo
- t
- replace: true
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.over1k_part2
-
- Stage: Stage-3
- Stats-Aggr Operator
-
PREHOOK: query: insert overwrite table over1k_part2 partition(ds="foo",t) select si,i,b,f,t from over1k where t is null or t=27 order by i
PREHOOK: type: QUERY
PREHOOK: Input: default@over1k
Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/filter_join_breaktask.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/filter_join_breaktask.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/tez/filter_join_breaktask.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/tez/filter_join_breaktask.q.out Mon Oct 6 03:44:13 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/insert_values_tmp_table.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/insert_values_tmp_table.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/insert_values_tmp_table.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/insert_values_tmp_table.q.out Mon Oct 6 03:44:13 2014
@@ -8,15 +8,13 @@ POSTHOOK: Output: database:default
POSTHOOK: Output: default@acid_ivtt
PREHOOK: query: insert into table acid_ivtt values
(1, 109.23, 'mary had a little lamb'),
- (429496729, 0.14, 'its fleece was white as snow'),
- (-29496729, -0.14, 'negative values test')
+ (429496729, 0.14, 'its fleece was white as snow')
PREHOOK: type: QUERY
PREHOOK: Input: default@values__tmp__table__1
PREHOOK: Output: default@acid_ivtt
POSTHOOK: query: insert into table acid_ivtt values
(1, 109.23, 'mary had a little lamb'),
- (429496729, 0.14, 'its fleece was white as snow'),
- (-29496729, -0.14, 'negative values test')
+ (429496729, 0.14, 'its fleece was white as snow')
POSTHOOK: type: QUERY
POSTHOOK: Input: default@values__tmp__table__1
POSTHOOK: Output: default@acid_ivtt
@@ -31,6 +29,5 @@ POSTHOOK: query: select i, de, vc from a
POSTHOOK: type: QUERY
POSTHOOK: Input: default@acid_ivtt
#### A masked pattern was here ####
--29496729 -0.14 negative values test
1 109.23 mary had a little lamb
429496729 0.14 its fleece was white as snow
Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/join0.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/join0.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/join0.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/join0.q.out Mon Oct 6 03:44:13 2014
@@ -1,4 +1,4 @@
-Warning: Shuffle Join MERGEJOIN[15][tables = [src1, src2]] in Stage 'Reducer 2' is a cross product
+Warning: Shuffle Join JOIN[8][tables = [src1, src2]] in Stage 'Reducer 2' is a cross product
PREHOOK: query: EXPLAIN
SELECT src1.key as k1, src1.value as v1,
src2.key as k2, src2.value as v2 FROM
@@ -61,7 +61,7 @@ STAGE PLANS:
value expressions: _col0 (type: string), _col1 (type: string)
Reducer 2
Reduce Operator Tree:
- Merge Join Operator
+ Join Operator
condition map:
Inner Join 0 to 1
condition expressions:
@@ -97,7 +97,7 @@ STAGE PLANS:
Processor Tree:
ListSink
-Warning: Shuffle Join MERGEJOIN[15][tables = [src1, src2]] in Stage 'Reducer 2' is a cross product
+Warning: Shuffle Join JOIN[8][tables = [src1, src2]] in Stage 'Reducer 2' is a cross product
PREHOOK: query: EXPLAIN FORMATTED
SELECT src1.key as k1, src1.value as v1,
src2.key as k2, src2.value as v2 FROM
@@ -115,7 +115,7 @@ SELECT src1.key as k1, src1.value as v1,
SORT BY k1, v1, k2, v2
POSTHOOK: type: QUERY
#### A masked pattern was here ####
-Warning: Shuffle Join MERGEJOIN[15][tables = [src1, src2]] in Stage 'Reducer 2' is a cross product
+Warning: Shuffle Join JOIN[8][tables = [src1, src2]] in Stage 'Reducer 2' is a cross product
PREHOOK: query: SELECT src1.key as k1, src1.value as v1,
src2.key as k2, src2.value as v2 FROM
(SELECT * FROM src WHERE src.key < 10) src1
Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/join1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/join1.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/join1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/join1.q.out Mon Oct 6 03:44:13 2014
@@ -56,7 +56,7 @@ STAGE PLANS:
Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
Reducer 2
Reduce Operator Tree:
- Merge Join Operator
+ Join Operator
condition map:
Inner Join 0 to 1
condition expressions:
Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/limit_pushdown.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/limit_pushdown.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/limit_pushdown.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/limit_pushdown.q.out Mon Oct 6 03:44:13 2014
@@ -673,11 +673,9 @@ STAGE PLANS:
PREHOOK: query: select key,value from src order by key limit 0
PREHOOK: type: QUERY
-PREHOOK: Input: default@src
#### A masked pattern was here ####
POSTHOOK: query: select key,value from src order by key limit 0
POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
#### A masked pattern was here ####
PREHOOK: query: -- 2MR (applied to last RS)
explain
@@ -889,7 +887,7 @@ STAGE PLANS:
value expressions: _col1 (type: bigint)
Reducer 3
Reduce Operator Tree:
- Merge Join Operator
+ Join Operator
condition map:
Inner Join 0 to 1
condition expressions:
Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/mapjoin_decimal.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/mapjoin_decimal.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/mapjoin_decimal.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/mapjoin_decimal.q.out Mon Oct 6 03:44:13 2014
@@ -120,8 +120,6 @@ STAGE PLANS:
0 dec (type: decimal(4,2))
1 dec (type: decimal(4,0))
outputColumnNames: _col0, _col4
- input vertices:
- 1 Map 1
Statistics: Num rows: 577 Data size: 64680 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: decimal(4,2)), _col4 (type: decimal(4,0))
Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/mapjoin_mapjoin.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/mapjoin_mapjoin.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/tez/mapjoin_mapjoin.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/tez/mapjoin_mapjoin.q.out Mon Oct 6 03:44:13 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/metadata_only_queries.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/metadata_only_queries.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/metadata_only_queries.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/metadata_only_queries.q.out Mon Oct 6 03:44:13 2014
@@ -352,11 +352,9 @@ STAGE PLANS:
PREHOOK: query: select count(*), sum(1), sum(0.2), count(1), count(s), count(bo), count(bin), count(si) from stats_tbl
PREHOOK: type: QUERY
-PREHOOK: Input: default@stats_tbl
#### A masked pattern was here ####
POSTHOOK: query: select count(*), sum(1), sum(0.2), count(1), count(s), count(bo), count(bin), count(si) from stats_tbl
POSTHOOK: type: QUERY
-POSTHOOK: Input: default@stats_tbl
#### A masked pattern was here ####
9999 9999 1999.8 9999 9999 9999 9999 9999
PREHOOK: query: explain
@@ -377,11 +375,9 @@ STAGE PLANS:
PREHOOK: query: select min(i), max(i), min(b), max(b), min(f), max(f), min(d), max(d) from stats_tbl
PREHOOK: type: QUERY
-PREHOOK: Input: default@stats_tbl
#### A masked pattern was here ####
POSTHOOK: query: select min(i), max(i), min(b), max(b), min(f), max(f), min(d), max(d) from stats_tbl
POSTHOOK: type: QUERY
-POSTHOOK: Input: default@stats_tbl
#### A masked pattern was here ####
65536 65791 4294967296 4294967551 0.009999999776482582 99.9800033569336 0.01 50.0
PREHOOK: query: explain
@@ -402,11 +398,9 @@ STAGE PLANS:
PREHOOK: query: select count(*), sum(1), sum(0.2), count(1), count(s), count(bo), count(bin), count(si) from stats_tbl_part
PREHOOK: type: QUERY
-PREHOOK: Input: default@stats_tbl_part
#### A masked pattern was here ####
POSTHOOK: query: select count(*), sum(1), sum(0.2), count(1), count(s), count(bo), count(bin), count(si) from stats_tbl_part
POSTHOOK: type: QUERY
-POSTHOOK: Input: default@stats_tbl_part
#### A masked pattern was here ####
9489 9489 1897.8 9489 9489 9489 9489 9489
PREHOOK: query: explain
@@ -427,11 +421,9 @@ STAGE PLANS:
PREHOOK: query: select min(i), max(i), min(b), max(b), min(f), max(f), min(d), max(d) from stats_tbl_part
PREHOOK: type: QUERY
-PREHOOK: Input: default@stats_tbl_part
#### A masked pattern was here ####
POSTHOOK: query: select min(i), max(i), min(b), max(b), min(f), max(f), min(d), max(d) from stats_tbl_part
POSTHOOK: type: QUERY
-POSTHOOK: Input: default@stats_tbl_part
#### A masked pattern was here ####
65536 65791 4294967296 4294967551 0.009999999776482582 99.9800033569336 0.01 50.0
PREHOOK: query: explain select count(ts) from stats_tbl_part
Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/metadataonly1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/metadataonly1.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/tez/metadataonly1.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/tez/metadataonly1.q.out Mon Oct 6 03:44:13 2014 differ