You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by sz...@apache.org on 2015/01/09 08:41:19 UTC
svn commit: r1650453 [9/9] - in /hive/branches/spark: ./
common/src/java/org/apache/hadoop/hive/conf/
hbase-handler/src/java/org/apache/hadoop/hive/hbase/
hcatalog/server-extensions/src/main/java/org/apache/hive/hcatalog/listener/
hcatalog/server-exten...
Modified: hive/branches/spark/ql/src/test/results/clientpositive/infer_bucket_sort_grouping_operators.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/infer_bucket_sort_grouping_operators.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/infer_bucket_sort_grouping_operators.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/infer_bucket_sort_grouping_operators.q.out Fri Jan 9 07:41:17 2015
@@ -44,11 +44,11 @@ STAGE PLANS:
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: key (type: string), value (type: string)
- outputColumnNames: key, value
+ outputColumnNames: _col0, _col1
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: count(1)
- keys: key (type: string), value (type: string), '0' (type: string)
+ keys: _col0 (type: string), _col1 (type: string), '0' (type: string)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 1500 Data size: 15936 Basic stats: COMPLETE Column stats: NONE
@@ -231,11 +231,11 @@ STAGE PLANS:
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: key (type: string), value (type: string)
- outputColumnNames: key, value
+ outputColumnNames: _col0, _col1
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: count(1)
- keys: key (type: string), value (type: string), '0' (type: string)
+ keys: _col0 (type: string), _col1 (type: string), '0' (type: string)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
@@ -418,11 +418,11 @@ STAGE PLANS:
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: key (type: string), value (type: string)
- outputColumnNames: key, value
+ outputColumnNames: _col0, _col1
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: count(1)
- keys: key (type: string), value (type: string), '0' (type: string)
+ keys: _col0 (type: string), _col1 (type: string), '0' (type: string)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE
Modified: hive/branches/spark/ql/src/test/results/clientpositive/input23.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/input23.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/input23.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/input23.q.out Fri Jan 9 07:41:17 2015
@@ -129,7 +129,7 @@ STAGE PLANS:
name: default.srcpart
name: default.srcpart
Truncated Path -> Alias:
- /srcpart/ds=2008-04-08/hr=11 [$hdt$_0:$hdt$_0:a]
+ /srcpart/ds=2008-04-08/hr=11 [$hdt$_0:a]
Needs Tagging: true
Reduce Operator Tree:
Join Operator
Modified: hive/branches/spark/ql/src/test/results/clientpositive/join17.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/join17.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/join17.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/join17.q.out Fri Jan 9 07:41:17 2015
@@ -153,7 +153,7 @@ STAGE PLANS:
name: default.src
name: default.src
Truncated Path -> Alias:
- /src [$hdt$_0:$hdt$_0:src1, $hdt$_0:$hdt$_1:src1]
+ /src [$hdt$_0:src1, $hdt$_1:src1]
Needs Tagging: true
Reduce Operator Tree:
Join Operator
Modified: hive/branches/spark/ql/src/test/results/clientpositive/join32_lessSize.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/join32_lessSize.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/join32_lessSize.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/join32_lessSize.q.out Fri Jan 9 07:41:17 2015
@@ -1931,11 +1931,11 @@ STAGE PLANS:
Stage: Stage-9
Map Reduce Local Work
Alias -> Map Local Tables:
- $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:y
+ $hdt$_0:$hdt$_0:$hdt$_0:y
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:y
+ $hdt$_0:$hdt$_0:$hdt$_0:y
TableScan
alias: y
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
@@ -2089,12 +2089,12 @@ STAGE PLANS:
name: default.src1
name: default.src1
Truncated Path -> Alias:
- /src1 [$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_1:x]
+ /src1 [$hdt$_0:$hdt$_0:$hdt$_1:x]
Stage: Stage-8
Map Reduce Local Work
Alias -> Map Local Tables:
- $hdt$_0:$hdt$_1:z
+ $hdt$_1:z
Fetch Operator
limit: -1
Partition Description:
@@ -2144,7 +2144,7 @@ STAGE PLANS:
name: default.srcpart
name: default.srcpart
Alias -> Map Local Operator Tree:
- $hdt$_0:$hdt$_1:z
+ $hdt$_1:z
TableScan
alias: z
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Modified: hive/branches/spark/ql/src/test/results/clientpositive/join34.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/join34.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/join34.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/join34.q.out Fri Jan 9 07:41:17 2015
@@ -148,11 +148,11 @@ STAGE PLANS:
Stage: Stage-6
Map Reduce Local Work
Alias -> Map Local Tables:
- $hdt$_0:$hdt$_1:x
+ $hdt$_1:x
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- $hdt$_0:$hdt$_1:x
+ $hdt$_1:x
TableScan
alias: x
Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
@@ -374,7 +374,7 @@ STAGE PLANS:
name: default.src1
name: default.src1
Truncated Path -> Alias:
- /src [$hdt$_0:$hdt$_0-subquery1:$hdt$_0-subquery1:x, $hdt$_0:$hdt$_0-subquery2:$hdt$_0-subquery2:x]
+ /src [$hdt$_0-subquery1:$hdt$_0-subquery1:x, $hdt$_0-subquery2:$hdt$_0-subquery2:x]
Stage: Stage-0
Move Operator
Modified: hive/branches/spark/ql/src/test/results/clientpositive/join35.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/join35.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/join35.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/join35.q.out Fri Jan 9 07:41:17 2015
@@ -232,7 +232,7 @@ STAGE PLANS:
name: default.src
name: default.src
Truncated Path -> Alias:
- /src [$hdt$_0:$hdt$_0-subquery1:$hdt$_0-subquery1:$hdt$_0:$hdt$_0:x]
+ /src [$hdt$_0-subquery1:$hdt$_0-subquery1:$hdt$_0:$hdt$_0:x]
Needs Tagging: false
Reduce Operator Tree:
Group By Operator
@@ -262,11 +262,11 @@ STAGE PLANS:
Stage: Stage-7
Map Reduce Local Work
Alias -> Map Local Tables:
- $hdt$_0:$hdt$_1:x
+ $hdt$_1:x
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- $hdt$_0:$hdt$_1:x
+ $hdt$_1:x
TableScan
alias: x
Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
@@ -569,7 +569,7 @@ STAGE PLANS:
name: default.src
name: default.src
Truncated Path -> Alias:
- /src [$hdt$_0:$hdt$_0-subquery2:$hdt$_0-subquery2:$hdt$_0:$hdt$_0:x]
+ /src [$hdt$_0-subquery2:$hdt$_0-subquery2:$hdt$_0:$hdt$_0:x]
Needs Tagging: false
Reduce Operator Tree:
Group By Operator
Modified: hive/branches/spark/ql/src/test/results/clientpositive/join7.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/join7.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/join7.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/join7.q.out Fri Jan 9 07:41:17 2015
@@ -48,9 +48,8 @@ INSERT OVERWRITE TABLE dest1 SELECT c.c1
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
Stage-1 is a root stage
- Stage-2 depends on stages: Stage-1
- Stage-0 depends on stages: Stage-2
- Stage-3 depends on stages: Stage-0
+ Stage-0 depends on stages: Stage-1
+ Stage-2 depends on stages: Stage-0
STAGE PLANS:
Stage: Stage-1
@@ -88,32 +87,6 @@ STAGE PLANS:
Map-reduce partition columns: _col0 (type: string)
Statistics: Num rows: 55 Data size: 584 Basic stats: COMPLETE Column stats: NONE
value expressions: _col1 (type: string)
- Reduce Operator Tree:
- Join Operator
- condition map:
- Outer Join 0 to 1
- keys:
- 0 _col0 (type: string)
- 1 _col0 (type: string)
- outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 60 Data size: 642 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.SequenceFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
-
- Stage: Stage-2
- Map Reduce
- Map Operator Tree:
- TableScan
- Reduce Output Operator
- key expressions: _col0 (type: string)
- sort order: +
- Map-reduce partition columns: _col0 (type: string)
- Statistics: Num rows: 60 Data size: 642 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string)
TableScan
alias: src1
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
@@ -133,19 +106,21 @@ STAGE PLANS:
Reduce Operator Tree:
Join Operator
condition map:
- Left Outer Join0 to 1
+ Outer Join 0 to 1
+ Left Outer Join0 to 2
keys:
0 _col0 (type: string)
1 _col0 (type: string)
+ 2 _col0 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
- Statistics: Num rows: 66 Data size: 706 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 121 Data size: 1284 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int), _col3 (type: string), UDFToInteger(_col4) (type: int), _col5 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
- Statistics: Num rows: 66 Data size: 706 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 121 Data size: 1284 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 66 Data size: 706 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 121 Data size: 1284 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -162,7 +137,7 @@ STAGE PLANS:
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1
- Stage: Stage-3
+ Stage: Stage-2
Stats-Aggr Operator
PREHOOK: query: FROM (
@@ -208,7 +183,7 @@ POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
POSTHOOK: Output: default@dest1
POSTHOOK: Lineage: dest1.c1 EXPRESSION [(src)src1.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: dest1.c2 EXPRESSION [(src)src1.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: dest1.c2 SIMPLE [(src)src1.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.c3 EXPRESSION [(src)src1.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.c4 SIMPLE [(src)src1.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.c5 EXPRESSION [(src)src1.FieldSchema(name:key, type:string, comment:default), ]
@@ -221,9 +196,6 @@ POSTHOOK: query: SELECT dest1.* FROM des
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest1
#### A masked pattern was here ####
-NULL NULL 24 val_24 NULL NULL
-NULL NULL 24 val_24 NULL NULL
-NULL NULL 20 val_20 NULL NULL
11 val_11 NULL NULL NULL NULL
12 val_12 NULL NULL NULL NULL
12 val_12 NULL NULL NULL NULL
@@ -235,3 +207,6 @@ NULL NULL 20 val_20 NULL NULL
18 val_18 18 val_18 NULL NULL
18 val_18 18 val_18 NULL NULL
19 val_19 19 val_19 NULL NULL
+NULL NULL 20 val_20 NULL NULL
+NULL NULL 24 val_24 NULL NULL
+NULL NULL 24 val_24 NULL NULL
Modified: hive/branches/spark/ql/src/test/results/clientpositive/join9.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/join9.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/join9.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/join9.q.out Fri Jan 9 07:41:17 2015
@@ -214,8 +214,8 @@ STAGE PLANS:
name: default.srcpart
name: default.srcpart
Truncated Path -> Alias:
- /src [$hdt$_0:$hdt$_1:src2]
- /srcpart/ds=2008-04-08/hr=12 [$hdt$_0:$hdt$_0:src1]
+ /src [$hdt$_1:src2]
+ /srcpart/ds=2008-04-08/hr=12 [$hdt$_0:src1]
Needs Tagging: true
Reduce Operator Tree:
Join Operator
Modified: hive/branches/spark/ql/src/test/results/clientpositive/louter_join_ppr.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/louter_join_ppr.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/louter_join_ppr.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/louter_join_ppr.q.out Fri Jan 9 07:41:17 2015
@@ -761,11 +761,11 @@ STAGE PLANS:
name: default.srcpart
name: default.srcpart
Truncated Path -> Alias:
- /src [$hdt$_0:$hdt$_1:b]
- /srcpart/ds=2008-04-08/hr=11 [$hdt$_0:$hdt$_0:a]
- /srcpart/ds=2008-04-08/hr=12 [$hdt$_0:$hdt$_0:a]
- /srcpart/ds=2008-04-09/hr=11 [$hdt$_0:$hdt$_0:a]
- /srcpart/ds=2008-04-09/hr=12 [$hdt$_0:$hdt$_0:a]
+ /src [$hdt$_1:b]
+ /srcpart/ds=2008-04-08/hr=11 [$hdt$_0:a]
+ /srcpart/ds=2008-04-08/hr=12 [$hdt$_0:a]
+ /srcpart/ds=2008-04-09/hr=11 [$hdt$_0:a]
+ /srcpart/ds=2008-04-09/hr=12 [$hdt$_0:a]
Needs Tagging: true
Reduce Operator Tree:
Join Operator
@@ -1522,9 +1522,9 @@ STAGE PLANS:
name: default.srcpart
name: default.srcpart
Truncated Path -> Alias:
- /src [$hdt$_0:$hdt$_1:b]
- /srcpart/ds=2008-04-08/hr=11 [$hdt$_0:$hdt$_0:a]
- /srcpart/ds=2008-04-08/hr=12 [$hdt$_0:$hdt$_0:a]
+ /src [$hdt$_1:b]
+ /srcpart/ds=2008-04-08/hr=11 [$hdt$_0:a]
+ /srcpart/ds=2008-04-08/hr=12 [$hdt$_0:a]
Needs Tagging: true
Reduce Operator Tree:
Join Operator
Modified: hive/branches/spark/ql/src/test/results/clientpositive/mapjoin_mapjoin.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/mapjoin_mapjoin.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/mapjoin_mapjoin.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/mapjoin_mapjoin.q.out Fri Jan 9 07:41:17 2015
@@ -574,11 +574,11 @@ STAGE PLANS:
Stage: Stage-9
Map Reduce Local Work
Alias -> Map Local Tables:
- $hdt$_0:$hdt$_0:$hdt$_1:src
+ $hdt$_0:$hdt$_1:src
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- $hdt$_0:$hdt$_0:$hdt$_1:src
+ $hdt$_0:$hdt$_1:src
TableScan
alias: src
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
@@ -627,11 +627,11 @@ STAGE PLANS:
Stage: Stage-8
Map Reduce Local Work
Alias -> Map Local Tables:
- $hdt$_0:$hdt$_0:$hdt$_2:src
+ $hdt$_0:$hdt$_2:src
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- $hdt$_0:$hdt$_0:$hdt$_2:src
+ $hdt$_0:$hdt$_2:src
TableScan
alias: src
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Modified: hive/branches/spark/ql/src/test/results/clientpositive/metadataonly1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/metadataonly1.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/metadataonly1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/metadataonly1.q.out Fri Jan 9 07:41:17 2015
@@ -614,8 +614,8 @@ STAGE PLANS:
value expressions: _col0 (type: string)
auto parallelism: false
Path -> Alias:
- -mr-10005default.test1{ds=1} [$hdt$_0:$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:a2]
- -mr-10006default.test1{ds=2} [$hdt$_0:$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:a2]
+ -mr-10005default.test1{ds=1} [$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:a2]
+ -mr-10006default.test1{ds=2} [$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:a2]
Path -> Partition:
-mr-10005default.test1{ds=1}
Partition
@@ -696,8 +696,8 @@ STAGE PLANS:
name: default.test1
name: default.test1
Truncated Path -> Alias:
- -mr-10005default.test1{ds=1} [$hdt$_0:$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:a2]
- -mr-10006default.test1{ds=2} [$hdt$_0:$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:a2]
+ -mr-10005default.test1{ds=1} [$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:a2]
+ -mr-10006default.test1{ds=2} [$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:a2]
Needs Tagging: false
Reduce Operator Tree:
Group By Operator
@@ -858,8 +858,8 @@ STAGE PLANS:
name: default.test1
name: default.test1
Truncated Path -> Alias:
- /test1/ds=1 [$hdt$_0:$hdt$_0:$hdt$_0:a2]
- /test1/ds=2 [$hdt$_0:$hdt$_0:$hdt$_0:a2]
+ /test1/ds=1 [$hdt$_0:$hdt$_0:a2]
+ /test1/ds=2 [$hdt$_0:$hdt$_0:a2]
#### A masked pattern was here ####
Needs Tagging: true
Reduce Operator Tree:
Modified: hive/branches/spark/ql/src/test/results/clientpositive/multiMapJoin2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/multiMapJoin2.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/multiMapJoin2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/multiMapJoin2.q.out Fri Jan 9 07:41:17 2015
@@ -23,14 +23,14 @@ STAGE PLANS:
Stage: Stage-8
Map Reduce Local Work
Alias -> Map Local Tables:
- null-subquery1:$hdt$_0-subquery1:$hdt$_0:$hdt$_1:y1
+ null-subquery1:$hdt$_0-subquery1:$hdt$_1:y1
Fetch Operator
limit: -1
- null-subquery2:$hdt$_0-subquery2:$hdt$_0:$hdt$_1:y1
+ null-subquery2:$hdt$_0-subquery2:$hdt$_1:y1
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- null-subquery1:$hdt$_0-subquery1:$hdt$_0:$hdt$_1:y1
+ null-subquery1:$hdt$_0-subquery1:$hdt$_1:y1
TableScan
alias: y1
Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
@@ -45,7 +45,7 @@ STAGE PLANS:
keys:
0 _col0 (type: string)
1 _col0 (type: string)
- null-subquery2:$hdt$_0-subquery2:$hdt$_0:$hdt$_1:y1
+ null-subquery2:$hdt$_0-subquery2:$hdt$_1:y1
TableScan
alias: y1
Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
@@ -266,11 +266,11 @@ STAGE PLANS:
Stage: Stage-9
Map Reduce Local Work
Alias -> Map Local Tables:
- null-subquery2:$hdt$_0-subquery2:$hdt$_0:$hdt$_1:y1
+ null-subquery2:$hdt$_0-subquery2:$hdt$_1:y1
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- null-subquery2:$hdt$_0-subquery2:$hdt$_0:$hdt$_1:y1
+ null-subquery2:$hdt$_0-subquery2:$hdt$_1:y1
TableScan
alias: y1
Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
@@ -319,11 +319,11 @@ STAGE PLANS:
Stage: Stage-8
Map Reduce Local Work
Alias -> Map Local Tables:
- null-subquery1:$hdt$_0-subquery1:$hdt$_0:$hdt$_1:y1
+ null-subquery1:$hdt$_0-subquery1:$hdt$_1:y1
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- null-subquery1:$hdt$_0-subquery1:$hdt$_0:$hdt$_1:y1
+ null-subquery1:$hdt$_0-subquery1:$hdt$_1:y1
TableScan
alias: y1
Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
@@ -555,11 +555,11 @@ STAGE PLANS:
Stage: Stage-6
Map Reduce Local Work
Alias -> Map Local Tables:
- null-subquery2:$hdt$_0-subquery2:$hdt$_0:$hdt$_1:x1
+ null-subquery2:$hdt$_0-subquery2:$hdt$_1:x1
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- null-subquery2:$hdt$_0-subquery2:$hdt$_0:$hdt$_1:x1
+ null-subquery2:$hdt$_0-subquery2:$hdt$_1:x1
TableScan
alias: x1
Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
@@ -761,11 +761,11 @@ STAGE PLANS:
Stage: Stage-17
Map Reduce Local Work
Alias -> Map Local Tables:
- $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_1:y1
+ $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_1:y1
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_1:y1
+ $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_1:y1
TableScan
alias: y1
Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
@@ -993,11 +993,11 @@ STAGE PLANS:
Stage: Stage-18
Map Reduce Local Work
Alias -> Map Local Tables:
- $hdt$_0:$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:$hdt$_2:y1
+ $hdt$_0:$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_2:y1
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- $hdt$_0:$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:$hdt$_2:y1
+ $hdt$_0:$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_2:y1
TableScan
alias: y1
Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
@@ -1159,14 +1159,14 @@ STAGE PLANS:
Stage: Stage-9
Map Reduce Local Work
Alias -> Map Local Tables:
- $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_1:y1
+ $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_1:y1
Fetch Operator
limit: -1
- $hdt$_0:$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:$hdt$_2:y1
+ $hdt$_0:$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_2:y1
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_1:y1
+ $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_1:y1
TableScan
alias: y1
Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
@@ -1181,7 +1181,7 @@ STAGE PLANS:
keys:
0 _col0 (type: string)
1 _col0 (type: string)
- $hdt$_0:$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:$hdt$_2:y1
+ $hdt$_0:$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_2:y1
TableScan
alias: y1
Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
@@ -1658,11 +1658,11 @@ STAGE PLANS:
Stage: Stage-14
Map Reduce Local Work
Alias -> Map Local Tables:
- $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_1:x1
+ $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_1:x1
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_1:x1
+ $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_1:x1
TableScan
alias: x1
Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
@@ -1823,11 +1823,11 @@ STAGE PLANS:
Stage: Stage-7
Map Reduce Local Work
Alias -> Map Local Tables:
- $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_1:x1
+ $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_1:x1
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_1:x1
+ $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_1:x1
TableScan
alias: x1
Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
@@ -2222,11 +2222,11 @@ STAGE PLANS:
Stage: Stage-15
Map Reduce Local Work
Alias -> Map Local Tables:
- null-subquery1:$hdt$_0-subquery1:$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:$hdt$_1:a
+ null-subquery1:$hdt$_0-subquery1:$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:a
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- null-subquery1:$hdt$_0-subquery1:$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:$hdt$_1:a
+ null-subquery1:$hdt$_0-subquery1:$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:a
TableScan
alias: a
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
@@ -2382,11 +2382,11 @@ STAGE PLANS:
Stage: Stage-16
Map Reduce Local Work
Alias -> Map Local Tables:
- null-subquery2:$hdt$_0-subquery2:$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:$hdt$_1:a
+ null-subquery2:$hdt$_0-subquery2:$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:a
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- null-subquery2:$hdt$_0-subquery2:$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:$hdt$_1:a
+ null-subquery2:$hdt$_0-subquery2:$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:a
TableScan
alias: a
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Modified: hive/branches/spark/ql/src/test/results/clientpositive/ppd_join_filter.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/ppd_join_filter.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/ppd_join_filter.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/ppd_join_filter.q.out Fri Jan 9 07:41:17 2015
@@ -202,7 +202,7 @@ STAGE PLANS:
name: default.src
name: default.src
Truncated Path -> Alias:
- /src [$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:a]
+ /src [$hdt$_1:$hdt$_1:$hdt$_1:a]
Needs Tagging: false
Reduce Operator Tree:
Group By Operator
@@ -345,7 +345,7 @@ STAGE PLANS:
name: default.src
name: default.src
Truncated Path -> Alias:
- /src [$hdt$_0:$hdt$_0:a]
+ /src [$hdt$_0:a]
#### A masked pattern was here ####
Needs Tagging: true
Reduce Operator Tree:
@@ -627,7 +627,7 @@ STAGE PLANS:
name: default.src
name: default.src
Truncated Path -> Alias:
- /src [$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:a]
+ /src [$hdt$_1:$hdt$_1:$hdt$_1:a]
Needs Tagging: false
Reduce Operator Tree:
Group By Operator
@@ -762,7 +762,7 @@ STAGE PLANS:
name: default.src
name: default.src
Truncated Path -> Alias:
- /src [$hdt$_0:$hdt$_0:a]
+ /src [$hdt$_0:a]
#### A masked pattern was here ####
Needs Tagging: true
Reduce Operator Tree:
@@ -1040,7 +1040,7 @@ STAGE PLANS:
name: default.src
name: default.src
Truncated Path -> Alias:
- /src [$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:a]
+ /src [$hdt$_1:$hdt$_1:$hdt$_1:a]
Needs Tagging: false
Reduce Operator Tree:
Group By Operator
@@ -1179,7 +1179,7 @@ STAGE PLANS:
name: default.src
name: default.src
Truncated Path -> Alias:
- /src [$hdt$_0:$hdt$_0:a]
+ /src [$hdt$_0:a]
#### A masked pattern was here ####
Needs Tagging: true
Reduce Operator Tree:
@@ -1461,7 +1461,7 @@ STAGE PLANS:
name: default.src
name: default.src
Truncated Path -> Alias:
- /src [$hdt$_0:$hdt$_1:$hdt$_1:$hdt$_1:a]
+ /src [$hdt$_1:$hdt$_1:$hdt$_1:a]
Needs Tagging: false
Reduce Operator Tree:
Group By Operator
@@ -1596,7 +1596,7 @@ STAGE PLANS:
name: default.src
name: default.src
Truncated Path -> Alias:
- /src [$hdt$_0:$hdt$_0:a]
+ /src [$hdt$_0:a]
#### A masked pattern was here ####
Needs Tagging: true
Reduce Operator Tree:
Modified: hive/branches/spark/ql/src/test/results/clientpositive/ppd_union_view.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/ppd_union_view.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/ppd_union_view.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/ppd_union_view.q.out Fri Jan 9 07:41:17 2015
@@ -476,8 +476,8 @@ STAGE PLANS:
outputColumnNames: _col1, _col3
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
- expressions: _col3 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
+ expressions: _col3 (type: string), _col1 (type: string), '2011-10-15' (type: string)
+ outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
File Output Operator
compressed: false
@@ -488,8 +488,8 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
properties:
- columns _col0,_col1
- columns.types string,string
+ columns _col0,_col1,_col2
+ columns.types string,string,string
escape.delim \
serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
@@ -505,13 +505,13 @@ STAGE PLANS:
Statistics: Num rows: 1 Data size: 11 Basic stats: COMPLETE Column stats: NONE
GatherStats: false
Select Operator
- expressions: key (type: string), value (type: string)
- outputColumnNames: _col0, _col1
+ expressions: key (type: string), value (type: string), '2011-10-15' (type: string)
+ outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 1 Data size: 11 Basic stats: COMPLETE Column stats: NONE
Union
Statistics: Num rows: 1 Data size: 11 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col0 (type: string), _col1 (type: string), '2011-10-15' (type: string)
+ expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string)
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 1 Data size: 11 Basic stats: COMPLETE Column stats: NONE
File Output Operator
@@ -540,7 +540,7 @@ STAGE PLANS:
Union
Statistics: Num rows: 1 Data size: 11 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col0 (type: string), _col1 (type: string), '2011-10-15' (type: string)
+ expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string)
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 1 Data size: 11 Basic stats: COMPLETE Column stats: NONE
File Output Operator
@@ -573,8 +573,8 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
properties:
- columns _col0,_col1
- columns.types string,string
+ columns _col0,_col1,_col2
+ columns.types string,string,string
escape.delim \
serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
@@ -582,8 +582,8 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
properties:
- columns _col0,_col1
- columns.types string,string
+ columns _col0,_col1,_col2
+ columns.types string,string,string
escape.delim \
serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
Modified: hive/branches/spark/ql/src/test/results/clientpositive/ppd_vc.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/ppd_vc.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/ppd_vc.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/ppd_vc.q.out Fri Jan 9 07:41:17 2015
@@ -679,11 +679,11 @@ STAGE PLANS:
name: default.srcpart
name: default.srcpart
Truncated Path -> Alias:
- /src [$hdt$_0:$hdt$_0:a]
- /srcpart/ds=2008-04-08/hr=11 [$hdt$_0:$hdt$_1:$hdt$_1:srcpart]
- /srcpart/ds=2008-04-08/hr=12 [$hdt$_0:$hdt$_1:$hdt$_1:srcpart]
- /srcpart/ds=2008-04-09/hr=11 [$hdt$_0:$hdt$_1:$hdt$_1:srcpart]
- /srcpart/ds=2008-04-09/hr=12 [$hdt$_0:$hdt$_1:$hdt$_1:srcpart]
+ /src [$hdt$_0:a]
+ /srcpart/ds=2008-04-08/hr=11 [$hdt$_1:$hdt$_1:srcpart]
+ /srcpart/ds=2008-04-08/hr=12 [$hdt$_1:$hdt$_1:srcpart]
+ /srcpart/ds=2008-04-09/hr=11 [$hdt$_1:$hdt$_1:srcpart]
+ /srcpart/ds=2008-04-09/hr=12 [$hdt$_1:$hdt$_1:srcpart]
Needs Tagging: true
Reduce Operator Tree:
Join Operator
Modified: hive/branches/spark/ql/src/test/results/clientpositive/router_join_ppr.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/router_join_ppr.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/router_join_ppr.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/router_join_ppr.q.out Fri Jan 9 07:41:17 2015
@@ -772,9 +772,9 @@ STAGE PLANS:
name: default.srcpart
name: default.srcpart
Truncated Path -> Alias:
- /src [$hdt$_0:$hdt$_1:b]
- /srcpart/ds=2008-04-08/hr=11 [$hdt$_0:$hdt$_0:$hdt$_0:a]
- /srcpart/ds=2008-04-08/hr=12 [$hdt$_0:$hdt$_0:$hdt$_0:a]
+ /src [$hdt$_1:b]
+ /srcpart/ds=2008-04-08/hr=11 [$hdt$_0:$hdt$_0:a]
+ /srcpart/ds=2008-04-08/hr=12 [$hdt$_0:$hdt$_0:a]
Needs Tagging: true
Reduce Operator Tree:
Join Operator
@@ -1526,9 +1526,9 @@ STAGE PLANS:
name: default.srcpart
name: default.srcpart
Truncated Path -> Alias:
- /src [$hdt$_0:$hdt$_1:$hdt$_1:b]
- /srcpart/ds=2008-04-08/hr=11 [$hdt$_0:$hdt$_0:$hdt$_0:a]
- /srcpart/ds=2008-04-08/hr=12 [$hdt$_0:$hdt$_0:$hdt$_0:a]
+ /src [$hdt$_1:$hdt$_1:b]
+ /srcpart/ds=2008-04-08/hr=11 [$hdt$_0:$hdt$_0:a]
+ /srcpart/ds=2008-04-08/hr=12 [$hdt$_0:$hdt$_0:a]
Needs Tagging: true
Reduce Operator Tree:
Join Operator
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/auto_join7.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/auto_join7.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/auto_join7.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/auto_join7.q.out Fri Jan 9 07:41:17 2015
@@ -47,48 +47,42 @@ FROM (
INSERT OVERWRITE TABLE dest1 SELECT c.c1, c.c2, c.c3, c.c4, c.c5, c.c6
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-3 is a root stage
- Stage-1 depends on stages: Stage-3
+ Stage-1 is a root stage
Stage-0 depends on stages: Stage-1
Stage-2 depends on stages: Stage-0
STAGE PLANS:
- Stage: Stage-3
+ Stage: Stage-1
Spark
+ Edges:
+ Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 1), Map 3 (PARTITION-LEVEL SORT, 1), Map 4 (PARTITION-LEVEL SORT, 1)
#### A masked pattern was here ####
Vertices:
- Map 4
+ Map 1
Map Operator Tree:
TableScan
alias: src1
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Filter Operator
- predicate: ((UDFToDouble(key) > 20.0) and (UDFToDouble(key) < 25.0)) (type: boolean)
+ predicate: ((UDFToDouble(key) > 10.0) and (UDFToDouble(key) < 20.0)) (type: boolean)
Statistics: Num rows: 55 Data size: 584 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: key (type: string), value (type: string)
outputColumnNames: _col0, _col1
Statistics: Num rows: 55 Data size: 584 Basic stats: COMPLETE Column stats: NONE
- Spark HashTable Sink Operator
- keys:
- 0 _col0 (type: string)
- 1 _col0 (type: string)
- Local Work:
- Map Reduce Local Work
-
- Stage: Stage-1
- Spark
- Edges:
- Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 1), Map 3 (PARTITION-LEVEL SORT, 1)
-#### A masked pattern was here ####
- Vertices:
- Map 1
+ Reduce Output Operator
+ key expressions: _col0 (type: string)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 55 Data size: 584 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col1 (type: string)
+ Map 3
Map Operator Tree:
TableScan
alias: src1
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Filter Operator
- predicate: ((UDFToDouble(key) > 10.0) and (UDFToDouble(key) < 20.0)) (type: boolean)
+ predicate: ((UDFToDouble(key) > 15.0) and (UDFToDouble(key) < 25.0)) (type: boolean)
Statistics: Num rows: 55 Data size: 584 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: key (type: string), value (type: string)
@@ -100,13 +94,13 @@ STAGE PLANS:
Map-reduce partition columns: _col0 (type: string)
Statistics: Num rows: 55 Data size: 584 Basic stats: COMPLETE Column stats: NONE
value expressions: _col1 (type: string)
- Map 3
+ Map 4
Map Operator Tree:
TableScan
alias: src1
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Filter Operator
- predicate: ((UDFToDouble(key) > 15.0) and (UDFToDouble(key) < 25.0)) (type: boolean)
+ predicate: ((UDFToDouble(key) > 20.0) and (UDFToDouble(key) < 25.0)) (type: boolean)
Statistics: Num rows: 55 Data size: 584 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: key (type: string), value (type: string)
@@ -119,39 +113,29 @@ STAGE PLANS:
Statistics: Num rows: 55 Data size: 584 Basic stats: COMPLETE Column stats: NONE
value expressions: _col1 (type: string)
Reducer 2
- Local Work:
- Map Reduce Local Work
Reduce Operator Tree:
Join Operator
condition map:
Outer Join 0 to 1
+ Left Outer Join0 to 2
keys:
0 _col0 (type: string)
1 _col0 (type: string)
- outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 60 Data size: 642 Basic stats: COMPLETE Column stats: NONE
- Map Join Operator
- condition map:
- Left Outer Join0 to 1
- keys:
- 0 _col0 (type: string)
- 1 _col0 (type: string)
+ 2 _col0 (type: string)
+ outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
+ Statistics: Num rows: 121 Data size: 1284 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int), _col3 (type: string), UDFToInteger(_col4) (type: int), _col5 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
- input vertices:
- 1 Map 4
- Statistics: Num rows: 66 Data size: 706 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int), _col3 (type: string), UDFToInteger(_col4) (type: int), _col5 (type: string)
- outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
- Statistics: Num rows: 66 Data size: 706 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 66 Data size: 706 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.dest1
+ Statistics: Num rows: 121 Data size: 1284 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 121 Data size: 1284 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.dest1
Stage: Stage-0
Move Operator
@@ -209,7 +193,7 @@ POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
POSTHOOK: Output: default@dest1
POSTHOOK: Lineage: dest1.c1 EXPRESSION [(src)src1.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: dest1.c2 EXPRESSION [(src)src1.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: dest1.c2 SIMPLE [(src)src1.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.c3 EXPRESSION [(src)src1.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.c4 SIMPLE [(src)src1.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.c5 EXPRESSION [(src)src1.FieldSchema(name:key, type:string, comment:default), ]
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_cube1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_cube1.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_cube1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_cube1.q.out Fri Jan 9 07:41:17 2015
@@ -109,6 +109,96 @@ NULL 18 1
NULL 28 1
NULL NULL 6
PREHOOK: query: EXPLAIN
+SELECT key, val, GROUPING__ID, count(1) FROM T1 GROUP BY key, val with cube
+PREHOOK: type: QUERY
+POSTHOOK: query: EXPLAIN
+SELECT key, val, GROUPING__ID, count(1) FROM T1 GROUP BY key, val with cube
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+ Stage-1 is a root stage
+ Stage-0 depends on stages: Stage-1
+
+STAGE PLANS:
+ Stage: Stage-1
+ Spark
+ Edges:
+ Reducer 2 <- Map 1 (GROUP, 3)
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: t1
+ Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
+ Select Operator
+ expressions: key (type: string), val (type: string)
+ outputColumnNames: key, val
+ Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
+ Group By Operator
+ aggregations: count(1)
+ keys: key (type: string), val (type: string), '0' (type: string)
+ mode: hash
+ outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string)
+ sort order: +++
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: string)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+ value expressions: _col3 (type: bigint)
+ Reducer 2
+ Reduce Operator Tree:
+ Group By Operator
+ aggregations: count(VALUE._col0)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: string)
+ mode: mergepartial
+ outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: bigint)
+ outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+ Stage: Stage-0
+ Fetch Operator
+ limit: -1
+ Processor Tree:
+ ListSink
+
+PREHOOK: query: SELECT key, val, GROUPING__ID, count(1) FROM T1 GROUP BY key, val with cube
+PREHOOK: type: QUERY
+PREHOOK: Input: default@t1
+#### A masked pattern was here ####
+POSTHOOK: query: SELECT key, val, GROUPING__ID, count(1) FROM T1 GROUP BY key, val with cube
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@t1
+#### A masked pattern was here ####
+1 11 3 1
+1 NULL 1 1
+2 12 3 1
+2 NULL 1 1
+3 13 3 1
+3 NULL 1 1
+7 17 3 1
+7 NULL 1 1
+8 18 3 1
+8 28 3 1
+8 NULL 1 2
+NULL 11 2 1
+NULL 12 2 1
+NULL 13 2 1
+NULL 17 2 1
+NULL 18 2 1
+NULL 28 2 1
+NULL NULL 0 6
+PREHOOK: query: EXPLAIN
SELECT key, count(distinct val) FROM T1 GROUP BY key with cube
PREHOOK: type: QUERY
POSTHOOK: query: EXPLAIN
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/join7.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/join7.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/join7.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/join7.q.out Fri Jan 9 07:41:17 2015
@@ -55,8 +55,7 @@ STAGE PLANS:
Stage: Stage-1
Spark
Edges:
- Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 1), Map 4 (PARTITION-LEVEL SORT, 1)
- Reducer 3 <- Map 5 (PARTITION-LEVEL SORT, 1), Reducer 2 (PARTITION-LEVEL SORT, 1)
+ Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 1), Map 3 (PARTITION-LEVEL SORT, 1), Map 4 (PARTITION-LEVEL SORT, 1)
#### A masked pattern was here ####
Vertices:
Map 1
@@ -77,7 +76,7 @@ STAGE PLANS:
Map-reduce partition columns: _col0 (type: string)
Statistics: Num rows: 55 Data size: 584 Basic stats: COMPLETE Column stats: NONE
value expressions: _col1 (type: string)
- Map 4
+ Map 3
Map Operator Tree:
TableScan
alias: src1
@@ -95,7 +94,7 @@ STAGE PLANS:
Map-reduce partition columns: _col0 (type: string)
Statistics: Num rows: 55 Data size: 584 Basic stats: COMPLETE Column stats: NONE
value expressions: _col1 (type: string)
- Map 5
+ Map 4
Map Operator Tree:
TableScan
alias: src1
@@ -118,34 +117,20 @@ STAGE PLANS:
Join Operator
condition map:
Outer Join 0 to 1
+ Left Outer Join0 to 2
keys:
0 _col0 (type: string)
1 _col0 (type: string)
- outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 60 Data size: 642 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: string)
- sort order: +
- Map-reduce partition columns: _col0 (type: string)
- Statistics: Num rows: 60 Data size: 642 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string)
- Reducer 3
- Reduce Operator Tree:
- Join Operator
- condition map:
- Left Outer Join0 to 1
- keys:
- 0 _col0 (type: string)
- 1 _col0 (type: string)
+ 2 _col0 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
- Statistics: Num rows: 66 Data size: 706 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 121 Data size: 1284 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int), _col3 (type: string), UDFToInteger(_col4) (type: int), _col5 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
- Statistics: Num rows: 66 Data size: 706 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 121 Data size: 1284 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 66 Data size: 706 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 121 Data size: 1284 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -208,7 +193,7 @@ POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
POSTHOOK: Output: default@dest1
POSTHOOK: Lineage: dest1.c1 EXPRESSION [(src)src1.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: dest1.c2 EXPRESSION [(src)src1.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: dest1.c2 SIMPLE [(src)src1.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.c3 EXPRESSION [(src)src1.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.c4 SIMPLE [(src)src1.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.c5 EXPRESSION [(src)src1.FieldSchema(name:key, type:string, comment:default), ]
@@ -221,9 +206,6 @@ POSTHOOK: query: SELECT dest1.* FROM des
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest1
#### A masked pattern was here ####
-NULL NULL 20 val_20 NULL NULL
-NULL NULL 24 val_24 NULL NULL
-NULL NULL 24 val_24 NULL NULL
11 val_11 NULL NULL NULL NULL
12 val_12 NULL NULL NULL NULL
12 val_12 NULL NULL NULL NULL
@@ -235,3 +217,6 @@ NULL NULL 24 val_24 NULL NULL
18 val_18 18 val_18 NULL NULL
18 val_18 18 val_18 NULL NULL
19 val_19 19 val_19 NULL NULL
+NULL NULL 20 val_20 NULL NULL
+NULL NULL 24 val_24 NULL NULL
+NULL NULL 24 val_24 NULL NULL
Modified: hive/branches/spark/ql/src/test/results/clientpositive/subquery_views.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/subquery_views.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/subquery_views.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/subquery_views.q.out Fri Jan 9 07:41:17 2015
@@ -70,7 +70,7 @@ POSTHOOK: Input: default@src
POSTHOOK: Output: database:default
POSTHOOK: Output: default@cv2
Warning: Shuffle Join JOIN[21][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product
-Warning: Shuffle Join JOIN[51][tables = [$hdt$_1, $hdt$_2]] in Stage 'Stage-6:MAPRED' is a cross product
+Warning: Shuffle Join JOIN[50][tables = [$hdt$_1, $hdt$_2]] in Stage 'Stage-6:MAPRED' is a cross product
PREHOOK: query: explain
select *
from cv2 where cv2.key in (select key from cv2 c where c.key < '11')
@@ -383,7 +383,7 @@ STAGE PLANS:
ListSink
Warning: Shuffle Join JOIN[21][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product
-Warning: Shuffle Join JOIN[51][tables = [$hdt$_1, $hdt$_2]] in Stage 'Stage-6:MAPRED' is a cross product
+Warning: Shuffle Join JOIN[50][tables = [$hdt$_1, $hdt$_2]] in Stage 'Stage-6:MAPRED' is a cross product
PREHOOK: query: select *
from cv2 where cv2.key in (select key from cv2 c where c.key < '11')
PREHOOK: type: QUERY
Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_join.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_join.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_join.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_join.q.out Fri Jan 9 07:41:17 2015
@@ -1,4 +1,4 @@
-Warning: Shuffle Join MERGEJOIN[12][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product
+Warning: Shuffle Join MERGEJOIN[11][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product
PREHOOK: query: -- current
explain select src.key from src join src src2
PREHOOK: type: QUERY
@@ -67,7 +67,7 @@ STAGE PLANS:
Processor Tree:
ListSink
-Warning: Shuffle Join MERGEJOIN[12][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product
+Warning: Shuffle Join MERGEJOIN[11][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product
PREHOOK: query: -- ansi cross join
explain select src.key from src cross join src src2
PREHOOK: type: QUERY
Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_1.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_1.q.out Fri Jan 9 07:41:17 2015
@@ -24,7 +24,7 @@ POSTHOOK: type: CREATETABLE_AS_SELECT
POSTHOOK: Input: default@src
POSTHOOK: Output: database:default
POSTHOOK: Output: default@B
-Warning: Shuffle Join MERGEJOIN[12][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product
+Warning: Shuffle Join MERGEJOIN[11][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product
PREHOOK: query: explain select * from A join B
PREHOOK: type: QUERY
POSTHOOK: query: explain select * from A join B
Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_2.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_2.q.out Fri Jan 9 07:41:17 2015
@@ -24,7 +24,7 @@ POSTHOOK: type: CREATETABLE_AS_SELECT
POSTHOOK: Input: default@src
POSTHOOK: Output: database:default
POSTHOOK: Output: default@B
-Warning: Map Join MAPJOIN[12][bigTable=?] in task 'Map 1' is a cross product
+Warning: Map Join MAPJOIN[11][bigTable=?] in task 'Map 1' is a cross product
PREHOOK: query: explain select * from A join B
PREHOOK: type: QUERY
POSTHOOK: query: explain select * from A join B
Modified: hive/branches/spark/ql/src/test/results/clientpositive/vector_left_outer_join.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/vector_left_outer_join.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/vector_left_outer_join.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/vector_left_outer_join.q.out Fri Jan 9 07:41:17 2015
@@ -25,14 +25,14 @@ STAGE PLANS:
Stage: Stage-8
Map Reduce Local Work
Alias -> Map Local Tables:
- $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_1:c
+ $hdt$_0:$hdt$_1:c
Fetch Operator
limit: -1
- $hdt$_0:$hdt$_0:$hdt$_1:c
+ $hdt$_0:$hdt$_2:c
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_1:c
+ $hdt$_0:$hdt$_1:c
TableScan
alias: c
Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
@@ -44,7 +44,7 @@ STAGE PLANS:
keys:
0 _col1 (type: int)
1 _col0 (type: int)
- $hdt$_0:$hdt$_0:$hdt$_1:c
+ $hdt$_0:$hdt$_2:c
TableScan
alias: c
Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
Modified: hive/branches/spark/ql/src/test/results/clientpositive/vectorized_mapjoin.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/vectorized_mapjoin.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/vectorized_mapjoin.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/vectorized_mapjoin.q.out Fri Jan 9 07:41:17 2015
@@ -15,11 +15,11 @@ STAGE PLANS:
Stage: Stage-5
Map Reduce Local Work
Alias -> Map Local Tables:
- $hdt$_0:$hdt$_0:$hdt$_0:t1
+ $hdt$_0:$hdt$_0:t1
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- $hdt$_0:$hdt$_0:$hdt$_0:t1
+ $hdt$_0:$hdt$_0:t1
TableScan
alias: t1
Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
Modified: hive/branches/spark/ql/src/test/results/clientpositive/vectorized_nested_mapjoin.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/vectorized_nested_mapjoin.q.out?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/vectorized_nested_mapjoin.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/vectorized_nested_mapjoin.q.out Fri Jan 9 07:41:17 2015
@@ -11,14 +11,14 @@ STAGE PLANS:
Stage: Stage-8
Map Reduce Local Work
Alias -> Map Local Tables:
- $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:v1
+ $hdt$_0:$hdt$_0:$hdt$_0:v1
Fetch Operator
limit: -1
- $hdt$_0:$hdt$_0:$hdt$_1:v1
+ $hdt$_0:$hdt$_1:v1
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:v1
+ $hdt$_0:$hdt$_0:$hdt$_0:v1
TableScan
alias: v1
Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
@@ -33,7 +33,7 @@ STAGE PLANS:
keys:
0 _col0 (type: tinyint)
1 _col0 (type: tinyint)
- $hdt$_0:$hdt$_0:$hdt$_1:v1
+ $hdt$_0:$hdt$_1:v1
TableScan
alias: v1
Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
Modified: hive/branches/spark/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerDe.java
URL: http://svn.apache.org/viewvc/hive/branches/spark/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerDe.java?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerDe.java (original)
+++ hive/branches/spark/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerDe.java Fri Jan 9 07:41:17 2015
@@ -100,7 +100,7 @@ public class AvroSerDe extends AbstractS
|| properties.getProperty(AvroSerdeUtils.AvroTableProperties.SCHEMA_URL.getPropName()) != null
|| columnNameProperty == null || columnNameProperty.isEmpty()
|| columnTypeProperty == null || columnTypeProperty.isEmpty()) {
- schema = determineSchemaOrReturnErrorSchema(properties);
+ schema = determineSchemaOrReturnErrorSchema(configuration, properties);
} else {
// Get column names and sort order
columnNames = Arrays.asList(columnNameProperty.split(","));
@@ -159,10 +159,10 @@ public class AvroSerDe extends AbstractS
* any call, including calls to update the serde properties, meaning
* if the serde is in a bad state, there is no way to update that state.
*/
- public Schema determineSchemaOrReturnErrorSchema(Properties props) {
+ public Schema determineSchemaOrReturnErrorSchema(Configuration conf, Properties props) {
try {
configErrors = "";
- return AvroSerdeUtils.determineSchemaOrThrowException(props);
+ return AvroSerdeUtils.determineSchemaOrThrowException(conf, props);
} catch(AvroSerdeException he) {
LOG.warn("Encountered AvroSerdeException determining schema. Returning " +
"signal schema to indicate problem", he);
Modified: hive/branches/spark/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerdeUtils.java
URL: http://svn.apache.org/viewvc/hive/branches/spark/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerdeUtils.java?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerdeUtils.java (original)
+++ hive/branches/spark/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerdeUtils.java Fri Jan 9 07:41:17 2015
@@ -97,7 +97,7 @@ public class AvroSerdeUtils {
* @throws IOException if error while trying to read the schema from another location
* @throws AvroSerdeException if unable to find a schema or pointer to it in the properties
*/
- public static Schema determineSchemaOrThrowException(Properties properties)
+ public static Schema determineSchemaOrThrowException(Configuration conf, Properties properties)
throws IOException, AvroSerdeException {
String schemaString = properties.getProperty(AvroTableProperties.SCHEMA_LITERAL.getPropName());
if(schemaString != null && !schemaString.equals(SCHEMA_NONE))
@@ -109,7 +109,7 @@ public class AvroSerdeUtils {
throw new AvroSerdeException(EXCEPTION_MESSAGE);
try {
- Schema s = getSchemaFromFS(schemaString, new Configuration());
+ Schema s = getSchemaFromFS(schemaString, conf);
if (s == null) {
//in case schema is not a file system
return AvroSerdeUtils.getSchemaFor(new URL(schemaString).openStream());
Modified: hive/branches/spark/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestAvroSerdeUtils.java
URL: http://svn.apache.org/viewvc/hive/branches/spark/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestAvroSerdeUtils.java?rev=1650453&r1=1650452&r2=1650453&view=diff
==============================================================================
--- hive/branches/spark/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestAvroSerdeUtils.java (original)
+++ hive/branches/spark/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestAvroSerdeUtils.java Fri Jan 9 07:41:17 2015
@@ -117,26 +117,29 @@ public class TestAvroSerdeUtils {
@Test(expected=AvroSerdeException.class)
public void determineSchemaThrowsExceptionIfNoSchema() throws IOException, AvroSerdeException {
+ Configuration conf = new Configuration();
Properties prop = new Properties();
- AvroSerdeUtils.determineSchemaOrThrowException(prop);
+ AvroSerdeUtils.determineSchemaOrThrowException(conf, prop);
}
@Test
public void determineSchemaFindsLiterals() throws Exception {
String schema = TestAvroObjectInspectorGenerator.RECORD_SCHEMA;
+ Configuration conf = new Configuration();
Properties props = new Properties();
props.put(AvroSerdeUtils.SCHEMA_LITERAL, schema);
Schema expected = AvroSerdeUtils.getSchemaFor(schema);
- assertEquals(expected, AvroSerdeUtils.determineSchemaOrThrowException(props));
+ assertEquals(expected, AvroSerdeUtils.determineSchemaOrThrowException(conf, props));
}
@Test
public void detemineSchemaTriesToOpenUrl() throws AvroSerdeException, IOException {
+ Configuration conf = new Configuration();
Properties props = new Properties();
props.put(AvroSerdeUtils.SCHEMA_URL, "not:///a.real.url");
try {
- AvroSerdeUtils.determineSchemaOrThrowException(props);
+ AvroSerdeUtils.determineSchemaOrThrowException(conf, props);
fail("Should have tried to open that URL");
} catch (AvroSerdeException e) {
assertEquals("Unable to read schema from given path: not:///a.real.url", e.getMessage());
@@ -145,13 +148,14 @@ public class TestAvroSerdeUtils {
@Test
public void noneOptionWorksForSpecifyingSchemas() throws IOException, AvroSerdeException {
+ Configuration conf = new Configuration();
Properties props = new Properties();
// Combo 1: Both set to none
props.put(SCHEMA_URL, SCHEMA_NONE);
props.put(SCHEMA_LITERAL, SCHEMA_NONE);
try {
- determineSchemaOrThrowException(props);
+ determineSchemaOrThrowException(conf, props);
fail("Should have thrown exception with none set for both url and literal");
} catch(AvroSerdeException he) {
assertEquals(EXCEPTION_MESSAGE, he.getMessage());
@@ -161,7 +165,7 @@ public class TestAvroSerdeUtils {
props.put(SCHEMA_LITERAL, TestAvroObjectInspectorGenerator.RECORD_SCHEMA);
Schema s;
try {
- s = determineSchemaOrThrowException(props);
+ s = determineSchemaOrThrowException(conf, props);
assertNotNull(s);
assertEquals(AvroSerdeUtils.getSchemaFor(TestAvroObjectInspectorGenerator.RECORD_SCHEMA), s);
} catch(AvroSerdeException he) {
@@ -172,7 +176,7 @@ public class TestAvroSerdeUtils {
props.put(SCHEMA_LITERAL, SCHEMA_NONE);
props.put(SCHEMA_URL, "not:///a.real.url");
try {
- determineSchemaOrThrowException(props);
+ determineSchemaOrThrowException(conf, props);
fail("Should have tried to open that bogus URL");
} catch (AvroSerdeException e) {
assertEquals("Unable to read schema from given path: not:///a.real.url", e.getMessage());