You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by ha...@apache.org on 2013/07/09 11:07:43 UTC
svn commit: r1501145 [9/13] - in /hive/branches/vectorization: ./
beeline/src/java/org/apache/hive/beeline/
beeline/src/test/org/apache/hive/beeline/src/test/
cli/src/java/org/apache/hadoop/hive/cli/
common/src/java/org/apache/hadoop/hive/common/metric...
Modified: hive/branches/vectorization/ql/src/test/results/clientpositive/groupby_grouping_sets4.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/clientpositive/groupby_grouping_sets4.q.out?rev=1501145&r1=1501144&r2=1501145&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/clientpositive/groupby_grouping_sets4.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/clientpositive/groupby_grouping_sets4.q.out Tue Jul 9 09:07:35 2013
@@ -19,7 +19,7 @@ SELECT * FROM
(SELECT a, b, count(*) from T1 where a < 3 group by a, b with cube) subq1
join
(SELECT a, b, count(*) from T1 where a < 3 group by a, b with cube) subq2
-on subq1.a = subq2.a
+on subq1.a = subq2.a order by subq1.a, subq1.b, subq2.a, subq2.b
PREHOOK: type: QUERY
POSTHOOK: query: -- This tests that cubes and rollups work fine inside sub-queries.
EXPLAIN
@@ -27,15 +27,16 @@ SELECT * FROM
(SELECT a, b, count(*) from T1 where a < 3 group by a, b with cube) subq1
join
(SELECT a, b, count(*) from T1 where a < 3 group by a, b with cube) subq2
-on subq1.a = subq2.a
+on subq1.a = subq2.a order by subq1.a, subq1.b, subq2.a, subq2.b
POSTHOOK: type: QUERY
ABSTRACT SYNTAX TREE:
- (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME T1))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL a)) (TOK_SELEXPR (TOK_TABLE_OR_COL b)) (TOK_SELEXPR (TOK_FUNCTIONSTAR count))) (TOK_WHERE (< (TOK_TABLE_OR_COL a) 3)) (TOK_CUBE_GROUPBY (TOK_TABLE_OR_COL a) (TOK_TABLE_OR_COL b)))) subq1) (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME T1))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL a)) (TOK_SELEXPR (TOK_TABLE_OR_COL b)) (TOK_SELEXPR (TOK_FUNCTIONSTAR count))) (TOK_WHERE (< (TOK_TABLE_OR_COL a) 3)) (TOK_CUBE_GROUPBY (TOK_TABLE_OR_COL a) (TOK_TABLE_OR_COL b)))) subq2) (= (. (TOK_TABLE_OR_COL subq1) a) (. (TOK_TABLE_OR_COL subq2) a)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF))))
+ (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME T1))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL a)) (TOK_SELEXPR (TOK_TABLE_OR_COL b)) (TOK_SELEXPR (TOK_FUNCTIONSTAR count))) (TOK_WHERE (< (TOK_TABLE_OR_COL a) 3)) (TOK_CUBE_GROUPBY (TOK_TABLE_OR_COL a) (TOK_TABLE_OR_COL b)))) subq1) (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME T1))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL a)) (TOK_SELEXPR (TOK_TABLE_OR_COL b)) (TOK_SELEXPR (TOK_FUNCTIONSTAR count))) (TOK_WHERE (< (TOK_TABLE_OR_COL a) 3)) (TOK_CUBE_GROUPBY (TOK_TABLE_OR_COL a) (TOK_TABLE_OR_COL b)))) subq2) (= (. (TOK_TABLE_OR_COL subq1) a) (. (TOK_TABLE_OR_COL subq2) a)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (. (TOK_TABLE_OR_COL subq1) a)) (TOK_TABSORTCOLNAMEASC (. (TOK_TAB
LE_OR_COL subq1) b)) (TOK_TABSORTCOLNAMEASC (. (TOK_TABLE_OR_COL subq2) a)) (TOK_TABSORTCOLNAMEASC (. (TOK_TABLE_OR_COL subq2) b)))))
STAGE DEPENDENCIES:
Stage-1 is a root stage
- Stage-2 depends on stages: Stage-1, Stage-3
- Stage-3 is a root stage
+ Stage-2 depends on stages: Stage-1, Stage-4
+ Stage-3 depends on stages: Stage-2
+ Stage-4 is a root stage
Stage-0 is a root stage
STAGE PLANS:
@@ -184,12 +185,50 @@ STAGE PLANS:
compressed: false
GlobalTableId: 0
table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
Stage: Stage-3
Map Reduce
Alias -> Map Operator Tree:
+#### A masked pattern was here ####
+ Reduce Output Operator
+ key expressions:
+ expr: _col0
+ type: string
+ expr: _col1
+ type: string
+ expr: _col3
+ type: string
+ expr: _col4
+ type: string
+ sort order: ++++
+ tag: -1
+ value expressions:
+ expr: _col0
+ type: string
+ expr: _col1
+ type: string
+ expr: _col2
+ type: bigint
+ expr: _col3
+ type: string
+ expr: _col4
+ type: string
+ expr: _col5
+ type: bigint
+ Reduce Operator Tree:
+ Extract
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+
+ Stage: Stage-4
+ Map Reduce
+ Alias -> Map Operator Tree:
subq2:t1
TableScan
alias: t1
@@ -276,7 +315,7 @@ PREHOOK: query: SELECT * FROM
(SELECT a, b, count(*) from T1 where a < 3 group by a, b with cube) subq1
join
(SELECT a, b, count(*) from T1 where a < 3 group by a, b with cube) subq2
-on subq1.a = subq2.a
+on subq1.a = subq2.a order by subq1.a, subq1.b, subq2.a, subq2.b
PREHOOK: type: QUERY
PREHOOK: Input: default@t1
#### A masked pattern was here ####
@@ -284,7 +323,7 @@ POSTHOOK: query: SELECT * FROM
(SELECT a, b, count(*) from T1 where a < 3 group by a, b with cube) subq1
join
(SELECT a, b, count(*) from T1 where a < 3 group by a, b with cube) subq2
-on subq1.a = subq2.a
+on subq1.a = subq2.a order by subq1.a, subq1.b, subq2.a, subq2.b
POSTHOOK: type: QUERY
POSTHOOK: Input: default@t1
#### A masked pattern was here ####
@@ -308,7 +347,7 @@ SELECT * FROM
(SELECT a, b, count(*) from T1 where a < 3 group by a, b with cube) subq1
join
(SELECT a, b, count(*) from T1 where a < 3 group by a, b with cube) subq2
-on subq1.a = subq2.a
+on subq1.a = subq2.a order by subq1.a, subq1.b, subq2.a, subq2.b
PREHOOK: type: QUERY
POSTHOOK: query: -- Since 4 grouping sets would be generated for each sub-query, an additional MR job should be created
-- for each of them
@@ -317,17 +356,18 @@ SELECT * FROM
(SELECT a, b, count(*) from T1 where a < 3 group by a, b with cube) subq1
join
(SELECT a, b, count(*) from T1 where a < 3 group by a, b with cube) subq2
-on subq1.a = subq2.a
+on subq1.a = subq2.a order by subq1.a, subq1.b, subq2.a, subq2.b
POSTHOOK: type: QUERY
ABSTRACT SYNTAX TREE:
- (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME T1))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL a)) (TOK_SELEXPR (TOK_TABLE_OR_COL b)) (TOK_SELEXPR (TOK_FUNCTIONSTAR count))) (TOK_WHERE (< (TOK_TABLE_OR_COL a) 3)) (TOK_CUBE_GROUPBY (TOK_TABLE_OR_COL a) (TOK_TABLE_OR_COL b)))) subq1) (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME T1))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL a)) (TOK_SELEXPR (TOK_TABLE_OR_COL b)) (TOK_SELEXPR (TOK_FUNCTIONSTAR count))) (TOK_WHERE (< (TOK_TABLE_OR_COL a) 3)) (TOK_CUBE_GROUPBY (TOK_TABLE_OR_COL a) (TOK_TABLE_OR_COL b)))) subq2) (= (. (TOK_TABLE_OR_COL subq1) a) (. (TOK_TABLE_OR_COL subq2) a)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF))))
+ (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME T1))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL a)) (TOK_SELEXPR (TOK_TABLE_OR_COL b)) (TOK_SELEXPR (TOK_FUNCTIONSTAR count))) (TOK_WHERE (< (TOK_TABLE_OR_COL a) 3)) (TOK_CUBE_GROUPBY (TOK_TABLE_OR_COL a) (TOK_TABLE_OR_COL b)))) subq1) (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME T1))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL a)) (TOK_SELEXPR (TOK_TABLE_OR_COL b)) (TOK_SELEXPR (TOK_FUNCTIONSTAR count))) (TOK_WHERE (< (TOK_TABLE_OR_COL a) 3)) (TOK_CUBE_GROUPBY (TOK_TABLE_OR_COL a) (TOK_TABLE_OR_COL b)))) subq2) (= (. (TOK_TABLE_OR_COL subq1) a) (. (TOK_TABLE_OR_COL subq2) a)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (. (TOK_TABLE_OR_COL subq1) a)) (TOK_TABSORTCOLNAMEASC (. (TOK_TAB
LE_OR_COL subq1) b)) (TOK_TABSORTCOLNAMEASC (. (TOK_TABLE_OR_COL subq2) a)) (TOK_TABSORTCOLNAMEASC (. (TOK_TABLE_OR_COL subq2) b)))))
STAGE DEPENDENCIES:
Stage-1 is a root stage
Stage-2 depends on stages: Stage-1
- Stage-3 depends on stages: Stage-2, Stage-5
- Stage-4 is a root stage
- Stage-5 depends on stages: Stage-4
+ Stage-3 depends on stages: Stage-2, Stage-6
+ Stage-4 depends on stages: Stage-3
+ Stage-5 is a root stage
+ Stage-6 depends on stages: Stage-5
Stage-0 is a root stage
STAGE PLANS:
@@ -515,12 +555,50 @@ STAGE PLANS:
compressed: false
GlobalTableId: 0
table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
Stage: Stage-4
Map Reduce
Alias -> Map Operator Tree:
+#### A masked pattern was here ####
+ Reduce Output Operator
+ key expressions:
+ expr: _col0
+ type: string
+ expr: _col1
+ type: string
+ expr: _col3
+ type: string
+ expr: _col4
+ type: string
+ sort order: ++++
+ tag: -1
+ value expressions:
+ expr: _col0
+ type: string
+ expr: _col1
+ type: string
+ expr: _col2
+ type: bigint
+ expr: _col3
+ type: string
+ expr: _col4
+ type: string
+ expr: _col5
+ type: bigint
+ Reduce Operator Tree:
+ Extract
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+
+ Stage: Stage-5
+ Map Reduce
+ Alias -> Map Operator Tree:
subq2:t1
TableScan
alias: t1
@@ -583,7 +661,7 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
- Stage: Stage-5
+ Stage: Stage-6
Map Reduce
Alias -> Map Operator Tree:
#### A masked pattern was here ####
@@ -646,7 +724,7 @@ PREHOOK: query: SELECT * FROM
(SELECT a, b, count(*) from T1 where a < 3 group by a, b with cube) subq1
join
(SELECT a, b, count(*) from T1 where a < 3 group by a, b with cube) subq2
-on subq1.a = subq2.a
+on subq1.a = subq2.a order by subq1.a, subq1.b, subq2.a, subq2.b
PREHOOK: type: QUERY
PREHOOK: Input: default@t1
#### A masked pattern was here ####
@@ -654,7 +732,7 @@ POSTHOOK: query: SELECT * FROM
(SELECT a, b, count(*) from T1 where a < 3 group by a, b with cube) subq1
join
(SELECT a, b, count(*) from T1 where a < 3 group by a, b with cube) subq2
-on subq1.a = subq2.a
+on subq1.a = subq2.a order by subq1.a, subq1.b, subq2.a, subq2.b
POSTHOOK: type: QUERY
POSTHOOK: Input: default@t1
#### A masked pattern was here ####
Modified: hive/branches/vectorization/ql/src/test/results/clientpositive/groupby_multi_single_reducer.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/clientpositive/groupby_multi_single_reducer.q.out?rev=1501145&r1=1501144&r2=1501145&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/clientpositive/groupby_multi_single_reducer.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/clientpositive/groupby_multi_single_reducer.q.out Tue Jul 9 09:07:35 2013
@@ -83,7 +83,7 @@ STAGE PLANS:
type: boolean
Group By Operator
aggregations:
- expr: count(DISTINCT KEY._col1:1._col0)
+ expr: count(DISTINCT KEY._col1:0._col0)
expr: sum(KEY._col1:1._col0)
expr: sum(DISTINCT KEY._col1:1._col0)
expr: count(VALUE._col0)
@@ -120,7 +120,7 @@ STAGE PLANS:
type: boolean
Group By Operator
aggregations:
- expr: count(DISTINCT KEY._col1:1._col0)
+ expr: count(DISTINCT KEY._col1:0._col0)
expr: sum(KEY._col1:1._col0)
expr: sum(DISTINCT KEY._col1:1._col0)
expr: count(VALUE._col0)
@@ -153,7 +153,7 @@ STAGE PLANS:
name: default.dest_g3
Group By Operator
aggregations:
- expr: count(DISTINCT KEY._col1:1._col0)
+ expr: count(DISTINCT KEY._col1:0._col0)
expr: sum(KEY._col1:1._col0)
expr: sum(DISTINCT KEY._col1:1._col0)
expr: count(VALUE._col0)
@@ -445,7 +445,7 @@ STAGE PLANS:
type: boolean
Group By Operator
aggregations:
- expr: count(DISTINCT KEY._col1:1._col0)
+ expr: count(DISTINCT KEY._col1:0._col0)
expr: sum(KEY._col1:1._col0)
expr: sum(DISTINCT KEY._col1:1._col0)
expr: count(VALUE._col0)
@@ -482,7 +482,7 @@ STAGE PLANS:
type: boolean
Group By Operator
aggregations:
- expr: count(DISTINCT KEY._col1:1._col0)
+ expr: count(DISTINCT KEY._col1:0._col0)
expr: sum(KEY._col1:1._col0)
expr: sum(DISTINCT KEY._col1:1._col0)
expr: count(VALUE._col0)
@@ -515,7 +515,7 @@ STAGE PLANS:
name: default.dest_g3
Group By Operator
aggregations:
- expr: count(DISTINCT KEY._col1:1._col0)
+ expr: count(DISTINCT KEY._col1:0._col0)
expr: sum(KEY._col1:1._col0)
expr: sum(DISTINCT KEY._col1:1._col0)
expr: count(VALUE._col0)
Modified: hive/branches/vectorization/ql/src/test/results/clientpositive/groupby_sort_1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/clientpositive/groupby_sort_1.q.out?rev=1501145&r1=1501144&r2=1501145&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/clientpositive/groupby_sort_1.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/clientpositive/groupby_sort_1.q.out Tue Jul 9 09:07:35 2013
@@ -4524,6 +4524,7 @@ STAGE PLANS:
columns _col0,_col1,_col2,_col3,_col4
columns.types string:bigint:string:string:bigint
escape.delim \
+ hive.serialization.extend.nesting.levels true
serialization.format 1
TotalFiles: 1
GatherStats: false
Modified: hive/branches/vectorization/ql/src/test/results/clientpositive/groupby_sort_skew_1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/clientpositive/groupby_sort_skew_1.q.out?rev=1501145&r1=1501144&r2=1501145&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/clientpositive/groupby_sort_skew_1.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/clientpositive/groupby_sort_skew_1.q.out Tue Jul 9 09:07:35 2013
@@ -4934,6 +4934,7 @@ STAGE PLANS:
columns _col0,_col1,_col2,_col3,_col4
columns.types string:bigint:string:string:bigint
escape.delim \
+ hive.serialization.extend.nesting.levels true
serialization.format 1
TotalFiles: 1
GatherStats: false
Modified: hive/branches/vectorization/ql/src/test/results/clientpositive/input12.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/clientpositive/input12.q.out?rev=1501145&r1=1501144&r2=1501145&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/clientpositive/input12.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/clientpositive/input12.q.out Tue Jul 9 09:07:35 2013
@@ -525,10 +525,12 @@ POSTHOOK: Lineage: dest3 PARTITION(ds=20
169 val_169
PREHOOK: query: SELECT dest3.* FROM dest3
PREHOOK: type: QUERY
+PREHOOK: Input: default@dest3
PREHOOK: Input: default@dest3@ds=2008-04-08/hr=12
#### A masked pattern was here ####
POSTHOOK: query: SELECT dest3.* FROM dest3
POSTHOOK: type: QUERY
+POSTHOOK: Input: default@dest3
POSTHOOK: Input: default@dest3@ds=2008-04-08/hr=12
#### A masked pattern was here ####
POSTHOOK: Lineage: dest1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
Modified: hive/branches/vectorization/ql/src/test/results/clientpositive/input23.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/clientpositive/input23.q.out?rev=1501145&r1=1501144&r2=1501145&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/clientpositive/input23.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/clientpositive/input23.q.out Tue Jul 9 09:07:35 2013
@@ -144,6 +144,7 @@ STAGE PLANS:
columns _col0,_col1,_col2,_col3,_col4,_col5,_col6,_col7
columns.types string:string:string:string:string:string:string:string
escape.delim \
+ hive.serialization.extend.nesting.levels true
serialization.format 1
TotalFiles: 1
GatherStats: false
Modified: hive/branches/vectorization/ql/src/test/results/clientpositive/input39.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/clientpositive/input39.q.out?rev=1501145&r1=1501144&r2=1501145&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/clientpositive/input39.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/clientpositive/input39.q.out Tue Jul 9 09:07:35 2013
@@ -96,9 +96,6 @@ STAGE PLANS:
expr: key
type: string
tag: 0
- value expressions:
- expr: ds
- type: string
t2
TableScan
alias: t2
@@ -115,18 +112,14 @@ STAGE PLANS:
expr: key
type: string
tag: 1
- value expressions:
- expr: ds
- type: string
Reduce Operator Tree:
Join Operator
condition map:
Inner Join 0 to 1
condition expressions:
- 0 {VALUE._col2}
- 1 {VALUE._col2}
+ 0
+ 1
handleSkewJoin: false
- outputColumnNames: _col2, _col7
Select Operator
Group By Operator
aggregations:
@@ -177,12 +170,16 @@ STAGE PLANS:
PREHOOK: query: select count(1) from t1 join t2 on t1.key=t2.key where t1.ds='1' and t2.ds='1'
PREHOOK: type: QUERY
+PREHOOK: Input: default@t1
PREHOOK: Input: default@t1@ds=1
+PREHOOK: Input: default@t2
PREHOOK: Input: default@t2@ds=1
#### A masked pattern was here ####
POSTHOOK: query: select count(1) from t1 join t2 on t1.key=t2.key where t1.ds='1' and t2.ds='1'
POSTHOOK: type: QUERY
+POSTHOOK: Input: default@t1
POSTHOOK: Input: default@t1@ds=1
+POSTHOOK: Input: default@t2
POSTHOOK: Input: default@t2@ds=1
#### A masked pattern was here ####
POSTHOOK: Lineage: t1 PARTITION(ds=1).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
Modified: hive/branches/vectorization/ql/src/test/results/clientpositive/input42.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/clientpositive/input42.q.out?rev=1501145&r1=1501144&r2=1501145&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/clientpositive/input42.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/clientpositive/input42.q.out Tue Jul 9 09:07:35 2013
@@ -158,6 +158,7 @@ STAGE PLANS:
columns _col0,_col1,_col2,_col3
columns.types string:string:string:string
escape.delim \
+ hive.serialization.extend.nesting.levels true
serialization.format 1
TotalFiles: 1
GatherStats: false
@@ -1348,6 +1349,7 @@ STAGE PLANS:
columns _col0,_col1,_col2,_col3
columns.types string:string:string:string
escape.delim \
+ hive.serialization.extend.nesting.levels true
serialization.format 1
TotalFiles: 1
GatherStats: false
@@ -1916,6 +1918,7 @@ STAGE PLANS:
columns _col0,_col1,_col2,_col3
columns.types string:string:string:string
escape.delim \
+ hive.serialization.extend.nesting.levels true
serialization.format 1
TotalFiles: 1
GatherStats: false
Modified: hive/branches/vectorization/ql/src/test/results/clientpositive/input_part10_win.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/clientpositive/input_part10_win.q.out?rev=1501145&r1=1501144&r2=1501145&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/clientpositive/input_part10_win.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/clientpositive/input_part10_win.q.out Tue Jul 9 09:07:35 2013
@@ -1,4 +1,5 @@
PREHOOK: query: -- INCLUDE_OS_WINDOWS
+-- included only on windows because of difference in file name encoding logic
CREATE TABLE part_special (
a STRING,
@@ -9,6 +10,7 @@ CREATE TABLE part_special (
)
PREHOOK: type: CREATETABLE
POSTHOOK: query: -- INCLUDE_OS_WINDOWS
+-- included only on windows because of difference in file name encoding logic
CREATE TABLE part_special (
a STRING,
@@ -105,18 +107,26 @@ POSTHOOK: query: DESCRIBE EXTENDED part_
POSTHOOK: type: DESCTABLE
POSTHOOK: Lineage: part_special PARTITION(ds=2008 04 08,ts=10:11:12=455).a SIMPLE []
POSTHOOK: Lineage: part_special PARTITION(ds=2008 04 08,ts=10:11:12=455).b SIMPLE []
-a string
-b string
-ds string
-ts string
+a string None
+b string None
+ds string None
+ts string None
+
+# Partition Information
+# col_name data_type comment
+
+ds string None
+ts string None
#### A masked pattern was here ####
PREHOOK: query: SELECT * FROM part_special WHERE ds='2008 04 08' AND ts = '10:11:12=455'
PREHOOK: type: QUERY
+PREHOOK: Input: default@part_special
PREHOOK: Input: default@part_special@ds=2008%2004%2008/ts=10%3A11%3A12%3D455
#### A masked pattern was here ####
POSTHOOK: query: SELECT * FROM part_special WHERE ds='2008 04 08' AND ts = '10:11:12=455'
POSTHOOK: type: QUERY
+POSTHOOK: Input: default@part_special
POSTHOOK: Input: default@part_special@ds=2008%2004%2008/ts=10%3A11%3A12%3D455
#### A masked pattern was here ####
POSTHOOK: Lineage: part_special PARTITION(ds=2008 04 08,ts=10:11:12=455).a SIMPLE []
Modified: hive/branches/vectorization/ql/src/test/results/clientpositive/input_part7.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/clientpositive/input_part7.q.out?rev=1501145&r1=1501144&r2=1501145&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/clientpositive/input_part7.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/clientpositive/input_part7.q.out Tue Jul 9 09:07:35 2013
@@ -242,6 +242,7 @@ STAGE PLANS:
columns _col0,_col1,_col2,_col3
columns.types string:string:string:string
escape.delim \
+ hive.serialization.extend.nesting.levels true
serialization.format 1
TotalFiles: 1
GatherStats: false
Modified: hive/branches/vectorization/ql/src/test/results/clientpositive/input_part9.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/clientpositive/input_part9.q.out?rev=1501145&r1=1501144&r2=1501145&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/clientpositive/input_part9.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/clientpositive/input_part9.q.out Tue Jul 9 09:07:35 2013
@@ -163,6 +163,7 @@ STAGE PLANS:
columns _col0,_col1,_col2,_col3
columns.types string:string:string:string
escape.delim \
+ hive.serialization.extend.nesting.levels true
serialization.format 1
TotalFiles: 1
GatherStats: false
Modified: hive/branches/vectorization/ql/src/test/results/clientpositive/join32_lessSize.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/clientpositive/join32_lessSize.q.out?rev=1501145&r1=1501144&r2=1501145&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/clientpositive/join32_lessSize.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/clientpositive/join32_lessSize.q.out Tue Jul 9 09:07:35 2013
@@ -375,11 +375,11 @@ POSTHOOK: Output: default@dest_j1
POSTHOOK: Lineage: dest_j1.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: dest_j1.val2 EXPRESSION [(src)y.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest_j1.value SIMPLE [(srcpart)z.FieldSchema(name:value, type:string, comment:default), ]
-PREHOOK: query: select * from dest_j1 x order by x.key
+PREHOOK: query: select * from dest_j1 x order by x.value
PREHOOK: type: QUERY
PREHOOK: Input: default@dest_j1
#### A masked pattern was here ####
-POSTHOOK: query: select * from dest_j1 x order by x.key
+POSTHOOK: query: select * from dest_j1 x order by x.value
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest_j1
#### A masked pattern was here ####
@@ -1540,11 +1540,11 @@ POSTHOOK: Lineage: dest_j1.val2 EXPRESSI
POSTHOOK: Lineage: dest_j1.val2 SIMPLE [(src)y.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest_j1.value SIMPLE [(srcpart)z.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest_j1.value SIMPLE [(src1)z.FieldSchema(name:value, type:string, comment:default), ]
-PREHOOK: query: select * from dest_j1 x order by x.key
+PREHOOK: query: select * from dest_j1 x order by x.value
PREHOOK: type: QUERY
PREHOOK: Input: default@dest_j1
#### A masked pattern was here ####
-POSTHOOK: query: select * from dest_j1 x order by x.key
+POSTHOOK: query: select * from dest_j1 x order by x.value
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest_j1
#### A masked pattern was here ####
@@ -2021,11 +2021,11 @@ POSTHOOK: Lineage: dest_j1.value SIMPLE
POSTHOOK: Lineage: dest_j2.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: dest_j2.val2 SIMPLE [(src1)x.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest_j2.value SIMPLE [(srcpart)z.FieldSchema(name:value, type:string, comment:default), ]
-PREHOOK: query: select * from dest_j2 x order by x.key
+PREHOOK: query: select * from dest_j2 x order by x.value
PREHOOK: type: QUERY
PREHOOK: Input: default@dest_j2
#### A masked pattern was here ####
-POSTHOOK: query: select * from dest_j2 x order by x.key
+POSTHOOK: query: select * from dest_j2 x order by x.value
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest_j2
#### A masked pattern was here ####
@@ -2521,11 +2521,11 @@ POSTHOOK: Lineage: dest_j2.val2 SIMPLE [
POSTHOOK: Lineage: dest_j2.val2 SIMPLE [(src1)x.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest_j2.value SIMPLE [(srcpart)z.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest_j2.value SIMPLE [(srcpart)z.FieldSchema(name:value, type:string, comment:default), ]
-PREHOOK: query: select * from dest_j2 x order by x.key
+PREHOOK: query: select * from dest_j2 x order by x.value
PREHOOK: type: QUERY
PREHOOK: Input: default@dest_j2
#### A masked pattern was here ####
-POSTHOOK: query: select * from dest_j2 x order by x.key
+POSTHOOK: query: select * from dest_j2 x order by x.value
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest_j2
#### A masked pattern was here ####
@@ -2541,23 +2541,16 @@ POSTHOOK: Lineage: dest_j2.val2 SIMPLE [
POSTHOOK: Lineage: dest_j2.val2 SIMPLE [(src1)x.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest_j2.value SIMPLE [(srcpart)z.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest_j2.value SIMPLE [(srcpart)z.FieldSchema(name:value, type:string, comment:default), ]
- val_27 val_27
- val_165 val_165
- val_409 val_409
- val_409 val_409
- val_409 val_409
- val_165 val_165
- val_484 val_484
- val_265 val_265
- val_265 val_265
- val_193 val_193
- val_193 val_193
- val_193 val_193
146 val_146 val_146
146 val_146 val_146
146 val_146 val_146
146 val_146 val_146
150 val_150 val_150
+ val_165 val_165
+ val_165 val_165
+ val_193 val_193
+ val_193 val_193
+ val_193 val_193
213 val_213 val_213
213 val_213 val_213
213 val_213 val_213
@@ -2570,6 +2563,9 @@ POSTHOOK: Lineage: dest_j2.value SIMPLE
255 val_255 val_255
255 val_255 val_255
255 val_255 val_255
+ val_265 val_265
+ val_265 val_265
+ val_27 val_27
273 val_273 val_273
273 val_273 val_273
273 val_273 val_273
@@ -2633,6 +2629,10 @@ POSTHOOK: Lineage: dest_j2.value SIMPLE
406 val_406 val_406
406 val_406 val_406
406 val_406 val_406
+ val_409 val_409
+ val_409 val_409
+ val_409 val_409
+ val_484 val_484
66 val_66 val_66
98 val_98 val_98
98 val_98 val_98
@@ -2833,11 +2833,11 @@ POSTHOOK: Lineage: dest_j2.val2 SIMPLE [
POSTHOOK: Lineage: dest_j2.value SIMPLE [(srcpart)z.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest_j2.value SIMPLE [(srcpart)z.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest_j2.value SIMPLE [(srcpart)x.FieldSchema(name:value, type:string, comment:default), ]
-PREHOOK: query: select * from dest_j2 x order by x.key
+PREHOOK: query: select * from dest_j2 x order by x.value
PREHOOK: type: QUERY
PREHOOK: Input: default@dest_j2
#### A masked pattern was here ####
-POSTHOOK: query: select * from dest_j2 x order by x.key
+POSTHOOK: query: select * from dest_j2 x order by x.value
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest_j2
#### A masked pattern was here ####
@@ -3142,11 +3142,11 @@ POSTHOOK: Lineage: dest_j2.value SIMPLE
POSTHOOK: Lineage: dest_j2.value SIMPLE [(srcpart)z.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest_j2.value SIMPLE [(srcpart)x.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest_j2.value SIMPLE [(srcpart)y.FieldSchema(name:value, type:string, comment:default), ]
-PREHOOK: query: select * from dest_j2 x order by x.key
+PREHOOK: query: select * from dest_j2 x order by x.value
PREHOOK: type: QUERY
PREHOOK: Input: default@dest_j2
#### A masked pattern was here ####
-POSTHOOK: query: select * from dest_j2 x order by x.key
+POSTHOOK: query: select * from dest_j2 x order by x.value
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest_j2
#### A masked pattern was here ####
Modified: hive/branches/vectorization/ql/src/test/results/clientpositive/join_1to1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/clientpositive/join_1to1.q.out?rev=1501145&r1=1501144&r2=1501145&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/clientpositive/join_1to1.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/clientpositive/join_1to1.q.out Tue Jul 9 09:07:35 2013
@@ -944,8 +944,6 @@ NULL 10050 66 NULL NULL NULL
50 10050 66 50 10050 66
50 10050 88 NULL NULL NULL
50 10050 88 NULL NULL NULL
-50 10050 88 NULL NULL NULL
-50 10050 88 NULL NULL NULL
60 10040 66 60 10040 66
60 10040 66 60 10040 66
60 10040 66 60 10040 66
@@ -1036,8 +1034,6 @@ NULL 10050 66 NULL NULL NULL
50 10050 66 50 10050 66
50 10050 88 NULL NULL NULL
50 10050 88 NULL NULL NULL
-50 10050 88 NULL NULL NULL
-50 10050 88 NULL NULL NULL
60 10040 66 60 10040 66
60 10040 66 60 10040 66
60 10040 66 60 10040 66
@@ -1162,28 +1158,18 @@ NULL 10050 66 NULL NULL NULL
25 10025 88 NULL NULL NULL
30 10030 66 NULL NULL NULL
35 10035 88 NULL NULL NULL
-40 10040 66 NULL NULL NULL
40 10040 66 40 10040 66
40 10040 88 NULL NULL NULL
-40 10040 88 NULL NULL NULL
50 10050 66 50 10050 66
50 10050 66 50 10050 66
50 10050 88 NULL NULL NULL
50 10050 88 NULL NULL NULL
-50 10050 88 NULL NULL NULL
-50 10050 88 NULL NULL NULL
-50 10050 88 NULL NULL NULL
-50 10050 88 NULL NULL NULL
60 10040 66 60 10040 66
60 10040 66 60 10040 66
60 10040 66 60 10040 66
60 10040 66 60 10040 66
70 10040 66 NULL NULL NULL
70 10040 66 NULL NULL NULL
-70 10040 66 NULL NULL NULL
-70 10040 66 NULL NULL NULL
-80 10040 88 NULL NULL NULL
-80 10040 88 NULL NULL NULL
80 10040 88 NULL NULL NULL
80 10040 88 NULL NULL NULL
PREHOOK: query: SELECT * FROM join_1to1_1 a full outer join join_1to1_2 b on a.key1 = b.key1 and a.key2 = b.key2 ORDER BY a.key1 ASC, a.key2 ASC, a.value ASC, b.key1 ASC, b.key2 ASC, b.value ASC
@@ -1262,27 +1248,17 @@ NULL 10050 66 NULL NULL NULL
25 10025 88 NULL NULL NULL
30 10030 66 NULL NULL NULL
35 10035 88 NULL NULL NULL
-40 10040 66 NULL NULL NULL
40 10040 66 40 10040 66
40 10040 88 NULL NULL NULL
-40 10040 88 NULL NULL NULL
50 10050 66 50 10050 66
50 10050 66 50 10050 66
50 10050 88 NULL NULL NULL
50 10050 88 NULL NULL NULL
-50 10050 88 NULL NULL NULL
-50 10050 88 NULL NULL NULL
-50 10050 88 NULL NULL NULL
-50 10050 88 NULL NULL NULL
60 10040 66 60 10040 66
60 10040 66 60 10040 66
60 10040 66 60 10040 66
60 10040 66 60 10040 66
70 10040 66 NULL NULL NULL
70 10040 66 NULL NULL NULL
-70 10040 66 NULL NULL NULL
-70 10040 66 NULL NULL NULL
-80 10040 88 NULL NULL NULL
-80 10040 88 NULL NULL NULL
80 10040 88 NULL NULL NULL
80 10040 88 NULL NULL NULL
Modified: hive/branches/vectorization/ql/src/test/results/clientpositive/join_filters_overlap.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/clientpositive/join_filters_overlap.q.out?rev=1501145&r1=1501144&r2=1501145&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/clientpositive/join_filters_overlap.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/clientpositive/join_filters_overlap.q.out Tue Jul 9 09:07:35 2013
@@ -181,6 +181,7 @@ STAGE PLANS:
columns _col0,_col1,_col2,_col3,_col4,_col5
columns.types int:int:int:int:int:int
escape.delim \
+ hive.serialization.extend.nesting.levels true
serialization.format 1
TotalFiles: 1
GatherStats: false
@@ -387,6 +388,7 @@ STAGE PLANS:
columns _col0,_col1,_col2,_col3,_col4,_col5
columns.types int:int:int:int:int:int
escape.delim \
+ hive.serialization.extend.nesting.levels true
serialization.format 1
TotalFiles: 1
GatherStats: false
@@ -593,6 +595,7 @@ STAGE PLANS:
columns _col0,_col1,_col2,_col3,_col4,_col5
columns.types int:int:int:int:int:int
escape.delim \
+ hive.serialization.extend.nesting.levels true
serialization.format 1
TotalFiles: 1
GatherStats: false
@@ -825,6 +828,7 @@ STAGE PLANS:
columns _col0,_col1,_col2,_col3,_col4,_col5,_col6,_col7
columns.types int:int:int:int:int:int:int:int
escape.delim \
+ hive.serialization.extend.nesting.levels true
serialization.format 1
TotalFiles: 1
GatherStats: false
@@ -1052,6 +1056,7 @@ STAGE PLANS:
columns _col0,_col1,_col2,_col3,_col4,_col5,_col6,_col7
columns.types int:int:int:int:int:int:int:int
escape.delim \
+ hive.serialization.extend.nesting.levels true
serialization.format 1
TotalFiles: 1
GatherStats: false
Modified: hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_1.q.out?rev=1501145&r1=1501144&r2=1501145&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_1.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_1.q.out Tue Jul 9 09:07:35 2013
@@ -423,6 +423,7 @@ STAGE PLANS:
columns _col0,_col1
columns.types string:string
escape.delim \
+ hive.serialization.extend.nesting.levels true
serialization.format 1
TotalFiles: 1
GatherStats: false
Modified: hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_11.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_11.q.out?rev=1501145&r1=1501144&r2=1501145&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_11.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_11.q.out Tue Jul 9 09:07:35 2013
@@ -279,6 +279,7 @@ STAGE PLANS:
columns _col0,_col1
columns.types string:string
escape.delim \
+ hive.serialization.extend.nesting.levels true
serialization.format 1
TotalFiles: 1
GatherStats: false
Modified: hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_12.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_12.q.out?rev=1501145&r1=1501144&r2=1501145&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_12.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_12.q.out Tue Jul 9 09:07:35 2013
@@ -247,11 +247,11 @@ Storage Desc Params:
serialization.format 1
PREHOOK: query: explain extended
select * from list_bucketing_mul_col
-where ds='2008-04-08' and hr='11' and col2 = "466" and col4 = "val_466"
+where ds='2008-04-08' and hr='11' and col2 = "466" and col4 = "val_466" ORDER BY col2, col4, ds, hr
PREHOOK: type: QUERY
POSTHOOK: query: explain extended
select * from list_bucketing_mul_col
-where ds='2008-04-08' and hr='11' and col2 = "466" and col4 = "val_466"
+where ds='2008-04-08' and hr='11' and col2 = "466" and col4 = "val_466" ORDER BY col2, col4, ds, hr
POSTHOOK: type: QUERY
POSTHOOK: Lineage: list_bucketing_mul_col PARTITION(ds=2008-04-08,hr=11).col1 EXPRESSION []
POSTHOOK: Lineage: list_bucketing_mul_col PARTITION(ds=2008-04-08,hr=11).col2 SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
@@ -259,7 +259,7 @@ POSTHOOK: Lineage: list_bucketing_mul_co
POSTHOOK: Lineage: list_bucketing_mul_col PARTITION(ds=2008-04-08,hr=11).col4 SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: list_bucketing_mul_col PARTITION(ds=2008-04-08,hr=11).col5 EXPRESSION []
ABSTRACT SYNTAX TREE:
- (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME list_bucketing_mul_col))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (and (and (and (= (TOK_TABLE_OR_COL ds) '2008-04-08') (= (TOK_TABLE_OR_COL hr) '11')) (= (TOK_TABLE_OR_COL col2) "466")) (= (TOK_TABLE_OR_COL col4) "val_466")))))
+ (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME list_bucketing_mul_col))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (and (and (and (= (TOK_TABLE_OR_COL ds) '2008-04-08') (= (TOK_TABLE_OR_COL hr) '11')) (= (TOK_TABLE_OR_COL col2) "466")) (= (TOK_TABLE_OR_COL col4) "val_466"))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL col2)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL col4)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL ds)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL hr)))))
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -295,23 +295,33 @@ STAGE PLANS:
expr: hr
type: string
outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6
- File Output Operator
- compressed: false
- GlobalTableId: 0
-#### A masked pattern was here ####
- NumFilesPerFileSink: 1
-#### A masked pattern was here ####
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- properties:
- columns _col0,_col1,_col2,_col3,_col4,_col5,_col6
- columns.types string:string:string:string:string:string:string
- escape.delim \
- serialization.format 1
- TotalFiles: 1
- GatherStats: false
- MultiFileSpray: false
+ Reduce Output Operator
+ key expressions:
+ expr: _col1
+ type: string
+ expr: _col3
+ type: string
+ expr: _col5
+ type: string
+ expr: _col6
+ type: string
+ sort order: ++++
+ tag: -1
+ value expressions:
+ expr: _col0
+ type: string
+ expr: _col1
+ type: string
+ expr: _col2
+ type: string
+ expr: _col3
+ type: string
+ expr: _col4
+ type: string
+ expr: _col5
+ type: string
+ expr: _col6
+ type: string
Needs Tagging: false
Path -> Alias:
#### A masked pattern was here ####
@@ -362,6 +372,26 @@ STAGE PLANS:
serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
name: default.list_bucketing_mul_col
name: default.list_bucketing_mul_col
+ Reduce Operator Tree:
+ Extract
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+#### A masked pattern was here ####
+ NumFilesPerFileSink: 1
+#### A masked pattern was here ####
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ columns _col0,_col1,_col2,_col3,_col4,_col5,_col6
+ columns.types string:string:string:string:string:string:string
+ escape.delim \
+ hive.serialization.extend.nesting.levels true
+ serialization.format 1
+ TotalFiles: 1
+ GatherStats: false
+ MultiFileSpray: false
Truncated Path -> Alias:
/list_bucketing_mul_col/ds=2008-04-08/hr=11/col2=466/col4=val_466 [list_bucketing_mul_col]
@@ -371,13 +401,13 @@ STAGE PLANS:
PREHOOK: query: select * from list_bucketing_mul_col
-where ds='2008-04-08' and hr='11' and col2 = "466" and col4 = "val_466"
+where ds='2008-04-08' and hr='11' and col2 = "466" and col4 = "val_466" ORDER BY col2, col4, ds, hr
PREHOOK: type: QUERY
PREHOOK: Input: default@list_bucketing_mul_col
PREHOOK: Input: default@list_bucketing_mul_col@ds=2008-04-08/hr=11
#### A masked pattern was here ####
POSTHOOK: query: select * from list_bucketing_mul_col
-where ds='2008-04-08' and hr='11' and col2 = "466" and col4 = "val_466"
+where ds='2008-04-08' and hr='11' and col2 = "466" and col4 = "val_466" ORDER BY col2, col4, ds, hr
POSTHOOK: type: QUERY
POSTHOOK: Input: default@list_bucketing_mul_col
POSTHOOK: Input: default@list_bucketing_mul_col@ds=2008-04-08/hr=11
@@ -392,11 +422,11 @@ POSTHOOK: Lineage: list_bucketing_mul_co
1 466 1 val_466 1 2008-04-08 11
PREHOOK: query: explain extended
select * from list_bucketing_mul_col
-where ds='2008-04-08' and hr='11' and col2 = "382" and col4 = "val_382"
+where ds='2008-04-08' and hr='11' and col2 = "382" and col4 = "val_382" ORDER BY col2, col4, ds, hr
PREHOOK: type: QUERY
POSTHOOK: query: explain extended
select * from list_bucketing_mul_col
-where ds='2008-04-08' and hr='11' and col2 = "382" and col4 = "val_382"
+where ds='2008-04-08' and hr='11' and col2 = "382" and col4 = "val_382" ORDER BY col2, col4, ds, hr
POSTHOOK: type: QUERY
POSTHOOK: Lineage: list_bucketing_mul_col PARTITION(ds=2008-04-08,hr=11).col1 EXPRESSION []
POSTHOOK: Lineage: list_bucketing_mul_col PARTITION(ds=2008-04-08,hr=11).col2 SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
@@ -404,7 +434,7 @@ POSTHOOK: Lineage: list_bucketing_mul_co
POSTHOOK: Lineage: list_bucketing_mul_col PARTITION(ds=2008-04-08,hr=11).col4 SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: list_bucketing_mul_col PARTITION(ds=2008-04-08,hr=11).col5 EXPRESSION []
ABSTRACT SYNTAX TREE:
- (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME list_bucketing_mul_col))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (and (and (and (= (TOK_TABLE_OR_COL ds) '2008-04-08') (= (TOK_TABLE_OR_COL hr) '11')) (= (TOK_TABLE_OR_COL col2) "382")) (= (TOK_TABLE_OR_COL col4) "val_382")))))
+ (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME list_bucketing_mul_col))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (and (and (and (= (TOK_TABLE_OR_COL ds) '2008-04-08') (= (TOK_TABLE_OR_COL hr) '11')) (= (TOK_TABLE_OR_COL col2) "382")) (= (TOK_TABLE_OR_COL col4) "val_382"))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL col2)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL col4)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL ds)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL hr)))))
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -440,23 +470,33 @@ STAGE PLANS:
expr: hr
type: string
outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6
- File Output Operator
- compressed: false
- GlobalTableId: 0
-#### A masked pattern was here ####
- NumFilesPerFileSink: 1
-#### A masked pattern was here ####
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- properties:
- columns _col0,_col1,_col2,_col3,_col4,_col5,_col6
- columns.types string:string:string:string:string:string:string
- escape.delim \
- serialization.format 1
- TotalFiles: 1
- GatherStats: false
- MultiFileSpray: false
+ Reduce Output Operator
+ key expressions:
+ expr: _col1
+ type: string
+ expr: _col3
+ type: string
+ expr: _col5
+ type: string
+ expr: _col6
+ type: string
+ sort order: ++++
+ tag: -1
+ value expressions:
+ expr: _col0
+ type: string
+ expr: _col1
+ type: string
+ expr: _col2
+ type: string
+ expr: _col3
+ type: string
+ expr: _col4
+ type: string
+ expr: _col5
+ type: string
+ expr: _col6
+ type: string
Needs Tagging: false
Path -> Alias:
#### A masked pattern was here ####
@@ -507,6 +547,26 @@ STAGE PLANS:
serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
name: default.list_bucketing_mul_col
name: default.list_bucketing_mul_col
+ Reduce Operator Tree:
+ Extract
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+#### A masked pattern was here ####
+ NumFilesPerFileSink: 1
+#### A masked pattern was here ####
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ columns _col0,_col1,_col2,_col3,_col4,_col5,_col6
+ columns.types string:string:string:string:string:string:string
+ escape.delim \
+ hive.serialization.extend.nesting.levels true
+ serialization.format 1
+ TotalFiles: 1
+ GatherStats: false
+ MultiFileSpray: false
Truncated Path -> Alias:
/list_bucketing_mul_col/ds=2008-04-08/hr=11/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME [list_bucketing_mul_col]
@@ -516,13 +576,13 @@ STAGE PLANS:
PREHOOK: query: select * from list_bucketing_mul_col
-where ds='2008-04-08' and hr='11' and col2 = "382" and col4 = "val_382"
+where ds='2008-04-08' and hr='11' and col2 = "382" and col4 = "val_382" ORDER BY col2, col4, ds, hr
PREHOOK: type: QUERY
PREHOOK: Input: default@list_bucketing_mul_col
PREHOOK: Input: default@list_bucketing_mul_col@ds=2008-04-08/hr=11
#### A masked pattern was here ####
POSTHOOK: query: select * from list_bucketing_mul_col
-where ds='2008-04-08' and hr='11' and col2 = "382" and col4 = "val_382"
+where ds='2008-04-08' and hr='11' and col2 = "382" and col4 = "val_382" ORDER BY col2, col4, ds, hr
POSTHOOK: type: QUERY
POSTHOOK: Input: default@list_bucketing_mul_col
POSTHOOK: Input: default@list_bucketing_mul_col@ds=2008-04-08/hr=11
Modified: hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_13.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_13.q.out?rev=1501145&r1=1501144&r2=1501145&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_13.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_13.q.out Tue Jul 9 09:07:35 2013
@@ -247,11 +247,11 @@ Storage Desc Params:
serialization.format 1
PREHOOK: query: explain extended
select * from list_bucketing_mul_col
-where ds='2008-04-08' and hr='2013-01-23+18:00:99' and col2 = "466" and col4 = "val_466"
+where ds='2008-04-08' and hr='2013-01-23+18:00:99' and col2 = "466" and col4 = "val_466" ORDER BY col2, col4, ds, hr
PREHOOK: type: QUERY
POSTHOOK: query: explain extended
select * from list_bucketing_mul_col
-where ds='2008-04-08' and hr='2013-01-23+18:00:99' and col2 = "466" and col4 = "val_466"
+where ds='2008-04-08' and hr='2013-01-23+18:00:99' and col2 = "466" and col4 = "val_466" ORDER BY col2, col4, ds, hr
POSTHOOK: type: QUERY
POSTHOOK: Lineage: list_bucketing_mul_col PARTITION(ds=2008-04-08,hr=2013-01-23+18:00:99).col1 EXPRESSION []
POSTHOOK: Lineage: list_bucketing_mul_col PARTITION(ds=2008-04-08,hr=2013-01-23+18:00:99).col2 SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
@@ -259,7 +259,7 @@ POSTHOOK: Lineage: list_bucketing_mul_co
POSTHOOK: Lineage: list_bucketing_mul_col PARTITION(ds=2008-04-08,hr=2013-01-23+18:00:99).col4 SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: list_bucketing_mul_col PARTITION(ds=2008-04-08,hr=2013-01-23+18:00:99).col5 EXPRESSION []
ABSTRACT SYNTAX TREE:
- (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME list_bucketing_mul_col))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (and (and (and (= (TOK_TABLE_OR_COL ds) '2008-04-08') (= (TOK_TABLE_OR_COL hr) '2013-01-23+18:00:99')) (= (TOK_TABLE_OR_COL col2) "466")) (= (TOK_TABLE_OR_COL col4) "val_466")))))
+ (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME list_bucketing_mul_col))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (and (and (and (= (TOK_TABLE_OR_COL ds) '2008-04-08') (= (TOK_TABLE_OR_COL hr) '2013-01-23+18:00:99')) (= (TOK_TABLE_OR_COL col2) "466")) (= (TOK_TABLE_OR_COL col4) "val_466"))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL col2)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL col4)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL ds)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL hr)))))
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -295,23 +295,33 @@ STAGE PLANS:
expr: hr
type: string
outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6
- File Output Operator
- compressed: false
- GlobalTableId: 0
-#### A masked pattern was here ####
- NumFilesPerFileSink: 1
-#### A masked pattern was here ####
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- properties:
- columns _col0,_col1,_col2,_col3,_col4,_col5,_col6
- columns.types string:string:string:string:string:string:string
- escape.delim \
- serialization.format 1
- TotalFiles: 1
- GatherStats: false
- MultiFileSpray: false
+ Reduce Output Operator
+ key expressions:
+ expr: _col1
+ type: string
+ expr: _col3
+ type: string
+ expr: _col5
+ type: string
+ expr: _col6
+ type: string
+ sort order: ++++
+ tag: -1
+ value expressions:
+ expr: _col0
+ type: string
+ expr: _col1
+ type: string
+ expr: _col2
+ type: string
+ expr: _col3
+ type: string
+ expr: _col4
+ type: string
+ expr: _col5
+ type: string
+ expr: _col6
+ type: string
Needs Tagging: false
Path -> Alias:
#### A masked pattern was here ####
@@ -362,6 +372,26 @@ STAGE PLANS:
serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
name: default.list_bucketing_mul_col
name: default.list_bucketing_mul_col
+ Reduce Operator Tree:
+ Extract
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+#### A masked pattern was here ####
+ NumFilesPerFileSink: 1
+#### A masked pattern was here ####
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ columns _col0,_col1,_col2,_col3,_col4,_col5,_col6
+ columns.types string:string:string:string:string:string:string
+ escape.delim \
+ hive.serialization.extend.nesting.levels true
+ serialization.format 1
+ TotalFiles: 1
+ GatherStats: false
+ MultiFileSpray: false
Truncated Path -> Alias:
/list_bucketing_mul_col/ds=2008-04-08/hr=2013-01-23+18%3A00%3A99/col2=466/col4=val_466 [list_bucketing_mul_col]
@@ -371,13 +401,13 @@ STAGE PLANS:
PREHOOK: query: select * from list_bucketing_mul_col
-where ds='2008-04-08' and hr='2013-01-23+18:00:99' and col2 = "466" and col4 = "val_466"
+where ds='2008-04-08' and hr='2013-01-23+18:00:99' and col2 = "466" and col4 = "val_466" ORDER BY col2, col4, ds, hr
PREHOOK: type: QUERY
PREHOOK: Input: default@list_bucketing_mul_col
PREHOOK: Input: default@list_bucketing_mul_col@ds=2008-04-08/hr=2013-01-23+18%3A00%3A99
#### A masked pattern was here ####
POSTHOOK: query: select * from list_bucketing_mul_col
-where ds='2008-04-08' and hr='2013-01-23+18:00:99' and col2 = "466" and col4 = "val_466"
+where ds='2008-04-08' and hr='2013-01-23+18:00:99' and col2 = "466" and col4 = "val_466" ORDER BY col2, col4, ds, hr
POSTHOOK: type: QUERY
POSTHOOK: Input: default@list_bucketing_mul_col
POSTHOOK: Input: default@list_bucketing_mul_col@ds=2008-04-08/hr=2013-01-23+18%3A00%3A99
Modified: hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_2.q.out?rev=1501145&r1=1501144&r2=1501145&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_2.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_2.q.out Tue Jul 9 09:07:35 2013
@@ -329,15 +329,15 @@ POSTHOOK: Lineage: list_bucketing_static
POSTHOOK: Lineage: list_bucketing_static_part PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
1000
PREHOOK: query: explain extended
-select * from list_bucketing_static_part where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484'
+select * from list_bucketing_static_part where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' ORDER BY key, value, ds, hr
PREHOOK: type: QUERY
POSTHOOK: query: explain extended
-select * from list_bucketing_static_part where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484'
+select * from list_bucketing_static_part where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' ORDER BY key, value, ds, hr
POSTHOOK: type: QUERY
POSTHOOK: Lineage: list_bucketing_static_part PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: list_bucketing_static_part PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
ABSTRACT SYNTAX TREE:
- (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME list_bucketing_static_part))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (and (and (and (= (TOK_TABLE_OR_COL ds) '2008-04-08') (= (TOK_TABLE_OR_COL hr) '11')) (= (TOK_TABLE_OR_COL key) '484')) (= (TOK_TABLE_OR_COL value) 'val_484')))))
+ (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME list_bucketing_static_part))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (and (and (and (= (TOK_TABLE_OR_COL ds) '2008-04-08') (= (TOK_TABLE_OR_COL hr) '11')) (= (TOK_TABLE_OR_COL key) '484')) (= (TOK_TABLE_OR_COL value) 'val_484'))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL value)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL ds)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL hr)))))
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -367,23 +367,27 @@ STAGE PLANS:
expr: hr
type: string
outputColumnNames: _col0, _col1, _col2, _col3
- File Output Operator
- compressed: false
- GlobalTableId: 0
-#### A masked pattern was here ####
- NumFilesPerFileSink: 1
-#### A masked pattern was here ####
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- properties:
- columns _col0,_col1,_col2,_col3
- columns.types string:string:string:string
- escape.delim \
- serialization.format 1
- TotalFiles: 1
- GatherStats: false
- MultiFileSpray: false
+ Reduce Output Operator
+ key expressions:
+ expr: _col0
+ type: string
+ expr: _col1
+ type: string
+ expr: _col2
+ type: string
+ expr: _col3
+ type: string
+ sort order: ++++
+ tag: -1
+ value expressions:
+ expr: _col0
+ type: string
+ expr: _col1
+ type: string
+ expr: _col2
+ type: string
+ expr: _col3
+ type: string
Needs Tagging: false
Path -> Alias:
#### A masked pattern was here ####
@@ -434,6 +438,26 @@ STAGE PLANS:
serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
name: default.list_bucketing_static_part
name: default.list_bucketing_static_part
+ Reduce Operator Tree:
+ Extract
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+#### A masked pattern was here ####
+ NumFilesPerFileSink: 1
+#### A masked pattern was here ####
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ columns _col0,_col1,_col2,_col3
+ columns.types string:string:string:string
+ escape.delim \
+ hive.serialization.extend.nesting.levels true
+ serialization.format 1
+ TotalFiles: 1
+ GatherStats: false
+ MultiFileSpray: false
Truncated Path -> Alias:
/list_bucketing_static_part/ds=2008-04-08/hr=11/key=484/value=val_484 [list_bucketing_static_part]
@@ -442,12 +466,12 @@ STAGE PLANS:
limit: -1
-PREHOOK: query: select * from list_bucketing_static_part where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484'
+PREHOOK: query: select * from list_bucketing_static_part where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' ORDER BY key, value, ds, hr
PREHOOK: type: QUERY
PREHOOK: Input: default@list_bucketing_static_part
PREHOOK: Input: default@list_bucketing_static_part@ds=2008-04-08/hr=11
#### A masked pattern was here ####
-POSTHOOK: query: select * from list_bucketing_static_part where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484'
+POSTHOOK: query: select * from list_bucketing_static_part where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' ORDER BY key, value, ds, hr
POSTHOOK: type: QUERY
POSTHOOK: Input: default@list_bucketing_static_part
POSTHOOK: Input: default@list_bucketing_static_part@ds=2008-04-08/hr=11
@@ -456,13 +480,13 @@ POSTHOOK: Lineage: list_bucketing_static
POSTHOOK: Lineage: list_bucketing_static_part PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
484 val_484 2008-04-08 11
484 val_484 2008-04-08 11
-PREHOOK: query: select * from srcpart where ds = '2008-04-08' and key = '484' and value = 'val_484'
+PREHOOK: query: select * from srcpart where ds = '2008-04-08' and key = '484' and value = 'val_484' ORDER BY key, value
PREHOOK: type: QUERY
PREHOOK: Input: default@srcpart
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
#### A masked pattern was here ####
-POSTHOOK: query: select * from srcpart where ds = '2008-04-08' and key = '484' and value = 'val_484'
+POSTHOOK: query: select * from srcpart where ds = '2008-04-08' and key = '484' and value = 'val_484' ORDER BY key, value
POSTHOOK: type: QUERY
POSTHOOK: Input: default@srcpart
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
@@ -470,11 +494,11 @@ POSTHOOK: Input: default@srcpart@ds=2008
#### A masked pattern was here ####
POSTHOOK: Lineage: list_bucketing_static_part PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: list_bucketing_static_part PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
-484 val_484 2008-04-08 12
484 val_484 2008-04-08 11
+484 val_484 2008-04-08 12
PREHOOK: query: -- 51 and val_51 in the table so skewed data for 51 and val_14 should be none
-- but query should succeed for 51 or 51 and val_14
-select * from srcpart where ds = '2008-04-08' and key = '51'
+select * from srcpart where ds = '2008-04-08' and key = '51' ORDER BY key, value
PREHOOK: type: QUERY
PREHOOK: Input: default@srcpart
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
@@ -482,7 +506,7 @@ PREHOOK: Input: default@srcpart@ds=2008-
#### A masked pattern was here ####
POSTHOOK: query: -- 51 and val_51 in the table so skewed data for 51 and val_14 should be none
-- but query should succeed for 51 or 51 and val_14
-select * from srcpart where ds = '2008-04-08' and key = '51'
+select * from srcpart where ds = '2008-04-08' and key = '51' ORDER BY key, value
POSTHOOK: type: QUERY
POSTHOOK: Input: default@srcpart
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
@@ -490,16 +514,16 @@ POSTHOOK: Input: default@srcpart@ds=2008
#### A masked pattern was here ####
POSTHOOK: Lineage: list_bucketing_static_part PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: list_bucketing_static_part PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
-51 val_51 2008-04-08 12
-51 val_51 2008-04-08 12
51 val_51 2008-04-08 11
51 val_51 2008-04-08 11
-PREHOOK: query: select * from list_bucketing_static_part where key = '51'
+51 val_51 2008-04-08 12
+51 val_51 2008-04-08 12
+PREHOOK: query: select * from list_bucketing_static_part where key = '51' ORDER BY key, value, ds, hr
PREHOOK: type: QUERY
PREHOOK: Input: default@list_bucketing_static_part
PREHOOK: Input: default@list_bucketing_static_part@ds=2008-04-08/hr=11
#### A masked pattern was here ####
-POSTHOOK: query: select * from list_bucketing_static_part where key = '51'
+POSTHOOK: query: select * from list_bucketing_static_part where key = '51' ORDER BY key, value, ds, hr
POSTHOOK: type: QUERY
POSTHOOK: Input: default@list_bucketing_static_part
POSTHOOK: Input: default@list_bucketing_static_part@ds=2008-04-08/hr=11
@@ -510,13 +534,13 @@ POSTHOOK: Lineage: list_bucketing_static
51 val_51 2008-04-08 11
51 val_51 2008-04-08 11
51 val_51 2008-04-08 11
-PREHOOK: query: select * from srcpart where ds = '2008-04-08' and key = '51' and value = 'val_14'
+PREHOOK: query: select * from srcpart where ds = '2008-04-08' and key = '51' and value = 'val_14' ORDER BY key, value
PREHOOK: type: QUERY
PREHOOK: Input: default@srcpart
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
#### A masked pattern was here ####
-POSTHOOK: query: select * from srcpart where ds = '2008-04-08' and key = '51' and value = 'val_14'
+POSTHOOK: query: select * from srcpart where ds = '2008-04-08' and key = '51' and value = 'val_14' ORDER BY key, value
POSTHOOK: type: QUERY
POSTHOOK: Input: default@srcpart
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
@@ -524,12 +548,12 @@ POSTHOOK: Input: default@srcpart@ds=2008
#### A masked pattern was here ####
POSTHOOK: Lineage: list_bucketing_static_part PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: list_bucketing_static_part PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
-PREHOOK: query: select * from list_bucketing_static_part where key = '51' and value = 'val_14'
+PREHOOK: query: select * from list_bucketing_static_part where key = '51' and value = 'val_14' ORDER BY key, value, ds, hr
PREHOOK: type: QUERY
PREHOOK: Input: default@list_bucketing_static_part
PREHOOK: Input: default@list_bucketing_static_part@ds=2008-04-08/hr=11
#### A masked pattern was here ####
-POSTHOOK: query: select * from list_bucketing_static_part where key = '51' and value = 'val_14'
+POSTHOOK: query: select * from list_bucketing_static_part where key = '51' and value = 'val_14' ORDER BY key, value, ds, hr
POSTHOOK: type: QUERY
POSTHOOK: Input: default@list_bucketing_static_part
POSTHOOK: Input: default@list_bucketing_static_part@ds=2008-04-08/hr=11
Modified: hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_3.q.out?rev=1501145&r1=1501144&r2=1501145&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_3.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_3.q.out Tue Jul 9 09:07:35 2013
@@ -352,6 +352,7 @@ STAGE PLANS:
columns _col0,_col1
columns.types string:string
escape.delim \
+ hive.serialization.extend.nesting.levels true
serialization.format 1
TotalFiles: 1
GatherStats: false
Modified: hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_4.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_4.q.out?rev=1501145&r1=1501144&r2=1501145&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_4.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/clientpositive/list_bucket_dml_4.q.out Tue Jul 9 09:07:35 2013
@@ -738,17 +738,17 @@ POSTHOOK: Lineage: list_bucketing_static
POSTHOOK: Lineage: list_bucketing_static_part PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
1000
PREHOOK: query: explain extended
-select * from list_bucketing_static_part where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484'
+select * from list_bucketing_static_part where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' ORDER BY key, value, ds, hr
PREHOOK: type: QUERY
POSTHOOK: query: explain extended
-select * from list_bucketing_static_part where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484'
+select * from list_bucketing_static_part where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' ORDER BY key, value, ds, hr
POSTHOOK: type: QUERY
POSTHOOK: Lineage: list_bucketing_static_part PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: list_bucketing_static_part PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: list_bucketing_static_part PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: list_bucketing_static_part PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
ABSTRACT SYNTAX TREE:
- (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME list_bucketing_static_part))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (and (and (and (= (TOK_TABLE_OR_COL ds) '2008-04-08') (= (TOK_TABLE_OR_COL hr) '11')) (= (TOK_TABLE_OR_COL key) '484')) (= (TOK_TABLE_OR_COL value) 'val_484')))))
+ (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME list_bucketing_static_part))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (and (and (and (= (TOK_TABLE_OR_COL ds) '2008-04-08') (= (TOK_TABLE_OR_COL hr) '11')) (= (TOK_TABLE_OR_COL key) '484')) (= (TOK_TABLE_OR_COL value) 'val_484'))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL value)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL ds)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL hr)))))
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -778,23 +778,27 @@ STAGE PLANS:
expr: hr
type: string
outputColumnNames: _col0, _col1, _col2, _col3
- File Output Operator
- compressed: false
- GlobalTableId: 0
-#### A masked pattern was here ####
- NumFilesPerFileSink: 1
-#### A masked pattern was here ####
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- properties:
- columns _col0,_col1,_col2,_col3
- columns.types string:string:string:string
- escape.delim \
- serialization.format 1
- TotalFiles: 1
- GatherStats: false
- MultiFileSpray: false
+ Reduce Output Operator
+ key expressions:
+ expr: _col0
+ type: string
+ expr: _col1
+ type: string
+ expr: _col2
+ type: string
+ expr: _col3
+ type: string
+ sort order: ++++
+ tag: -1
+ value expressions:
+ expr: _col0
+ type: string
+ expr: _col1
+ type: string
+ expr: _col2
+ type: string
+ expr: _col3
+ type: string
Needs Tagging: false
Path -> Alias:
#### A masked pattern was here ####
@@ -845,6 +849,26 @@ STAGE PLANS:
serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
name: default.list_bucketing_static_part
name: default.list_bucketing_static_part
+ Reduce Operator Tree:
+ Extract
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+#### A masked pattern was here ####
+ NumFilesPerFileSink: 1
+#### A masked pattern was here ####
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ columns _col0,_col1,_col2,_col3
+ columns.types string:string:string:string
+ escape.delim \
+ hive.serialization.extend.nesting.levels true
+ serialization.format 1
+ TotalFiles: 1
+ GatherStats: false
+ MultiFileSpray: false
Truncated Path -> Alias:
/list_bucketing_static_part/ds=2008-04-08/hr=11/key=484/value=val_484 [list_bucketing_static_part]
@@ -853,12 +877,12 @@ STAGE PLANS:
limit: -1
-PREHOOK: query: select * from list_bucketing_static_part where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484'
+PREHOOK: query: select * from list_bucketing_static_part where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' ORDER BY key, value, ds, hr
PREHOOK: type: QUERY
PREHOOK: Input: default@list_bucketing_static_part
PREHOOK: Input: default@list_bucketing_static_part@ds=2008-04-08/hr=11
#### A masked pattern was here ####
-POSTHOOK: query: select * from list_bucketing_static_part where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484'
+POSTHOOK: query: select * from list_bucketing_static_part where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' ORDER BY key, value, ds, hr
POSTHOOK: type: QUERY
POSTHOOK: Input: default@list_bucketing_static_part
POSTHOOK: Input: default@list_bucketing_static_part@ds=2008-04-08/hr=11
@@ -869,13 +893,13 @@ POSTHOOK: Lineage: list_bucketing_static
POSTHOOK: Lineage: list_bucketing_static_part PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
484 val_484 2008-04-08 11
484 val_484 2008-04-08 11
-PREHOOK: query: select * from srcpart where ds = '2008-04-08' and key = '484' and value = 'val_484'
+PREHOOK: query: select * from srcpart where ds = '2008-04-08' and key = '484' and value = 'val_484' ORDER BY key, value
PREHOOK: type: QUERY
PREHOOK: Input: default@srcpart
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
#### A masked pattern was here ####
-POSTHOOK: query: select * from srcpart where ds = '2008-04-08' and key = '484' and value = 'val_484'
+POSTHOOK: query: select * from srcpart where ds = '2008-04-08' and key = '484' and value = 'val_484' ORDER BY key, value
POSTHOOK: type: QUERY
POSTHOOK: Input: default@srcpart
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
@@ -885,8 +909,8 @@ POSTHOOK: Lineage: list_bucketing_static
POSTHOOK: Lineage: list_bucketing_static_part PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: list_bucketing_static_part PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: list_bucketing_static_part PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
-484 val_484 2008-04-08 12
484 val_484 2008-04-08 11
+484 val_484 2008-04-08 12
PREHOOK: query: -- clean up
drop table list_bucketing_static_part
PREHOOK: type: DROPTABLE