You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by se...@apache.org on 2015/01/23 20:59:24 UTC
svn commit: r1654355 [25/27] - in /hive/branches/llap: ./
beeline/src/java/org/apache/hive/beeline/
cli/src/java/org/apache/hadoop/hive/cli/
common/src/java/org/apache/hadoop/hive/common/
common/src/java/org/apache/hadoop/hive/conf/ data/conf/ data/con...
Modified: hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorized_ptf.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorized_ptf.q.out?rev=1654355&r1=1654354&r2=1654355&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorized_ptf.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorized_ptf.q.out Fri Jan 23 19:59:11 2015
@@ -268,7 +268,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -288,7 +288,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -312,12 +312,14 @@ STAGE PLANS:
Map-reduce partition columns: _col2 (type: string)
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
tag: -1
- value expressions: _col1 (type: string), _col2 (type: string), _col5 (type: int), _col7 (type: double)
+ value expressions: _col5 (type: int), _col7 (type: double)
auto parallelism: true
Reducer 3
Needs Tagging: false
Reduce Operator Tree:
- Extract
+ Select Operator
+ expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col3 (type: int), VALUE._col5 (type: double)
+ outputColumnNames: _col1, _col2, _col5, _col7
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
PTF Operator
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
@@ -558,7 +560,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -578,7 +580,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -625,7 +627,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -645,7 +647,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -690,12 +692,14 @@ STAGE PLANS:
Map-reduce partition columns: _col2 (type: string)
Statistics: Num rows: 14 Data size: 8823 Basic stats: COMPLETE Column stats: NONE
tag: -1
- value expressions: _col1 (type: string), _col2 (type: string), _col5 (type: int)
+ value expressions: _col5 (type: int)
auto parallelism: true
Reducer 4
Needs Tagging: false
Reduce Operator Tree:
- Extract
+ Select Operator
+ expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col3 (type: int)
+ outputColumnNames: _col1, _col2, _col5
Statistics: Num rows: 14 Data size: 8823 Basic stats: COMPLETE Column stats: NONE
PTF Operator
Statistics: Num rows: 14 Data size: 8823 Basic stats: COMPLETE Column stats: NONE
@@ -871,7 +875,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -891,7 +895,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -1128,7 +1132,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -1148,7 +1152,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -1172,12 +1176,14 @@ STAGE PLANS:
Map-reduce partition columns: _col2 (type: string)
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
tag: -1
- value expressions: _col1 (type: string), _col2 (type: string), _col5 (type: int), _col7 (type: double)
+ value expressions: _col5 (type: int), _col7 (type: double)
auto parallelism: true
Reducer 3
Needs Tagging: false
Reduce Operator Tree:
- Extract
+ Select Operator
+ expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col3 (type: int), VALUE._col5 (type: double)
+ outputColumnNames: _col1, _col2, _col5, _col7
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
PTF Operator
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
@@ -1416,7 +1422,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -1436,7 +1442,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -1460,12 +1466,14 @@ STAGE PLANS:
Map-reduce partition columns: _col2 (type: string)
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
tag: -1
- value expressions: _col1 (type: string), _col2 (type: string), _col5 (type: int)
+ value expressions: _col5 (type: int)
auto parallelism: true
Reducer 3
Needs Tagging: false
Reduce Operator Tree:
- Extract
+ Select Operator
+ expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col3 (type: int)
+ outputColumnNames: _col1, _col2, _col5
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
PTF Operator
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
@@ -1714,7 +1722,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -1734,7 +1742,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -1778,13 +1786,15 @@ STAGE PLANS:
Map-reduce partition columns: _col0 (type: string)
Statistics: Num rows: 13 Data size: 8021 Basic stats: COMPLETE Column stats: NONE
tag: -1
- value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ value expressions: _col2 (type: int)
auto parallelism: true
Execution mode: vectorized
Reducer 4
Needs Tagging: false
Reduce Operator Tree:
- Extract
+ Select Operator
+ expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), VALUE._col0 (type: int)
+ outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 13 Data size: 8021 Basic stats: COMPLETE Column stats: NONE
PTF Operator
Statistics: Num rows: 13 Data size: 8021 Basic stats: COMPLETE Column stats: NONE
@@ -1978,7 +1988,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -1998,7 +2008,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -2044,7 +2054,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -2064,7 +2074,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -2295,7 +2305,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -2315,7 +2325,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -2359,7 +2369,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -2379,7 +2389,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -2617,7 +2627,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -2637,7 +2647,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -2661,12 +2671,13 @@ STAGE PLANS:
Map-reduce partition columns: _col2 (type: string)
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
tag: -1
- value expressions: _col1 (type: string), _col2 (type: string), _col5 (type: int)
auto parallelism: true
Reducer 3
Needs Tagging: false
Reduce Operator Tree:
- Extract
+ Select Operator
+ expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), KEY.reducesinkkey2 (type: int)
+ outputColumnNames: _col1, _col2, _col5
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
PTF Operator
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
@@ -2894,7 +2905,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -2914,7 +2925,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -2938,12 +2949,14 @@ STAGE PLANS:
Map-reduce partition columns: _col2 (type: string)
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
tag: -1
- value expressions: _col1 (type: string), _col2 (type: string), _col5 (type: int), _col7 (type: double)
+ value expressions: _col5 (type: int), _col7 (type: double)
auto parallelism: true
Reducer 3
Needs Tagging: false
Reduce Operator Tree:
- Extract
+ Select Operator
+ expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col3 (type: int), VALUE._col5 (type: double)
+ outputColumnNames: _col1, _col2, _col5, _col7
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
PTF Operator
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
@@ -3173,7 +3186,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -3193,7 +3206,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -3217,12 +3230,14 @@ STAGE PLANS:
Map-reduce partition columns: _col2 (type: string)
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
tag: -1
- value expressions: _col1 (type: string), _col2 (type: string), _col5 (type: int), _col7 (type: double)
+ value expressions: _col5 (type: int), _col7 (type: double)
auto parallelism: true
Reducer 3
Needs Tagging: false
Reduce Operator Tree:
- Extract
+ Select Operator
+ expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col3 (type: int), VALUE._col5 (type: double)
+ outputColumnNames: _col1, _col2, _col5, _col7
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
PTF Operator
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
@@ -3462,7 +3477,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -3482,7 +3497,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -3523,12 +3538,14 @@ STAGE PLANS:
Map-reduce partition columns: _col2 (type: string)
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
tag: -1
- value expressions: _col1 (type: string), _col2 (type: string), _col5 (type: int), _col7 (type: double)
+ value expressions: _col5 (type: int), _col7 (type: double)
auto parallelism: true
Reducer 4
Needs Tagging: false
Reduce Operator Tree:
- Extract
+ Select Operator
+ expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col3 (type: int), VALUE._col5 (type: double)
+ outputColumnNames: _col1, _col2, _col5, _col7
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
PTF Operator
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
@@ -3788,7 +3805,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -3808,7 +3825,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -3832,12 +3849,14 @@ STAGE PLANS:
Map-reduce partition columns: _col2 (type: string)
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
tag: -1
- value expressions: _col1 (type: string), _col2 (type: string), _col5 (type: int), _col7 (type: double)
+ value expressions: _col5 (type: int), _col7 (type: double)
auto parallelism: true
Reducer 3
Needs Tagging: false
Reduce Operator Tree:
- Extract
+ Select Operator
+ expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col3 (type: int), VALUE._col5 (type: double)
+ outputColumnNames: _col1, _col2, _col5, _col7
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
PTF Operator
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
@@ -4174,7 +4193,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -4194,7 +4213,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -4240,7 +4259,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -4260,7 +4279,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -4309,12 +4328,14 @@ STAGE PLANS:
Map-reduce partition columns: _col2 (type: string)
Statistics: Num rows: 14 Data size: 8823 Basic stats: COMPLETE Column stats: NONE
tag: -1
- value expressions: _col1 (type: string), _col2 (type: string), _col5 (type: int), _col7 (type: double)
+ value expressions: _col5 (type: int), _col7 (type: double)
auto parallelism: true
Reducer 4
Needs Tagging: false
Reduce Operator Tree:
- Extract
+ Select Operator
+ expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col3 (type: int), VALUE._col5 (type: double)
+ outputColumnNames: _col1, _col2, _col5, _col7
Statistics: Num rows: 14 Data size: 8823 Basic stats: COMPLETE Column stats: NONE
PTF Operator
Statistics: Num rows: 14 Data size: 8823 Basic stats: COMPLETE Column stats: NONE
@@ -4501,7 +4522,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -4521,7 +4542,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -4783,7 +4804,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -4803,7 +4824,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -4842,12 +4863,14 @@ STAGE PLANS:
Map-reduce partition columns: _col0 (type: string)
Statistics: Num rows: 13 Data size: 8021 Basic stats: COMPLETE Column stats: NONE
tag: -1
- value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double)
+ value expressions: _col2 (type: double)
auto parallelism: true
Reducer 4
Needs Tagging: false
Reduce Operator Tree:
- Extract
+ Select Operator
+ expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), VALUE._col0 (type: double)
+ outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 13 Data size: 8021 Basic stats: COMPLETE Column stats: NONE
PTF Operator
Statistics: Num rows: 13 Data size: 8021 Basic stats: COMPLETE Column stats: NONE
@@ -5244,7 +5267,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -5264,7 +5287,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -5284,7 +5307,7 @@ STAGE PLANS:
Map-reduce partition columns: _col2 (type: string)
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
tag: -1
- value expressions: _col1 (type: string), _col2 (type: string), _col5 (type: int)
+ value expressions: _col1 (type: string)
auto parallelism: true
Select Operator
expressions: _col1 (type: string), _col2 (type: string), _col5 (type: int), _col7 (type: double)
@@ -5296,12 +5319,14 @@ STAGE PLANS:
Map-reduce partition columns: _col2 (type: string)
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
tag: -1
- value expressions: _col1 (type: string), _col2 (type: string), _col5 (type: int), _col7 (type: double)
+ value expressions: _col5 (type: int), _col7 (type: double)
auto parallelism: true
Reducer 3
Needs Tagging: false
Reduce Operator Tree:
- Extract
+ Select Operator
+ expressions: VALUE._col1 (type: string), KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: int)
+ outputColumnNames: _col1, _col2, _col5
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
PTF Operator
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
@@ -5311,12 +5336,14 @@ STAGE PLANS:
Map-reduce partition columns: _col2 (type: string)
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
tag: -1
- value expressions: _wcol0 (type: bigint), _col1 (type: string), _col2 (type: string), _col5 (type: int)
+ value expressions: _wcol0 (type: bigint), _col5 (type: int)
auto parallelism: true
Reducer 4
Needs Tagging: false
Reduce Operator Tree:
- Extract
+ Select Operator
+ expressions: VALUE._col0 (type: bigint), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col4 (type: int)
+ outputColumnNames: _col0, _col2, _col3, _col6
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
PTF Operator
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
@@ -5353,7 +5380,9 @@ STAGE PLANS:
Reducer 5
Needs Tagging: false
Reduce Operator Tree:
- Extract
+ Select Operator
+ expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col3 (type: int), VALUE._col5 (type: double)
+ outputColumnNames: _col1, _col2, _col5, _col7
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
PTF Operator
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
@@ -5752,7 +5781,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -5772,7 +5801,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -5813,12 +5842,14 @@ STAGE PLANS:
Map-reduce partition columns: _col2 (type: string), _col1 (type: string)
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
tag: -1
- value expressions: _col1 (type: string), _col2 (type: string), _col5 (type: int)
+ value expressions: _col5 (type: int)
auto parallelism: true
Reducer 4
Needs Tagging: false
Reduce Operator Tree:
- Extract
+ Select Operator
+ expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col3 (type: int)
+ outputColumnNames: _col1, _col2, _col5
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
PTF Operator
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
@@ -6109,7 +6140,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -6129,7 +6160,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -6183,12 +6214,14 @@ STAGE PLANS:
Map-reduce partition columns: _col2 (type: string)
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
tag: -1
- value expressions: _col1 (type: string), _col2 (type: string), _col5 (type: int)
+ value expressions: _col5 (type: int)
auto parallelism: true
Reducer 5
Needs Tagging: false
Reduce Operator Tree:
- Extract
+ Select Operator
+ expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col3 (type: int)
+ outputColumnNames: _col1, _col2, _col5
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
PTF Operator
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
@@ -6462,7 +6495,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -6482,7 +6515,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -6521,12 +6554,14 @@ STAGE PLANS:
Map-reduce partition columns: _col2 (type: string)
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
tag: -1
- value expressions: _col1 (type: string), _col2 (type: string), _col5 (type: int)
+ value expressions: _col5 (type: int)
auto parallelism: true
Reducer 4
Needs Tagging: false
Reduce Operator Tree:
- Extract
+ Select Operator
+ expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col3 (type: int)
+ outputColumnNames: _col1, _col2, _col5
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
PTF Operator
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
@@ -6812,7 +6847,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -6832,7 +6867,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -6888,12 +6923,14 @@ STAGE PLANS:
Map-reduce partition columns: _col2 (type: string), _col1 (type: string)
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
tag: -1
- value expressions: _col1 (type: string), _col2 (type: string), _col5 (type: int)
+ value expressions: _col5 (type: int)
auto parallelism: true
Reducer 5
Needs Tagging: false
Reduce Operator Tree:
- Extract
+ Select Operator
+ expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col3 (type: int)
+ outputColumnNames: _col1, _col2, _col5
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
PTF Operator
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
@@ -7210,7 +7247,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -7230,7 +7267,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -7271,12 +7308,14 @@ STAGE PLANS:
Map-reduce partition columns: _col2 (type: string), _col1 (type: string)
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
tag: -1
- value expressions: _col1 (type: string), _col2 (type: string), _col5 (type: int)
+ value expressions: _col5 (type: int)
auto parallelism: true
Reducer 4
Needs Tagging: false
Reduce Operator Tree:
- Extract
+ Select Operator
+ expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col3 (type: int)
+ outputColumnNames: _col1, _col2, _col5
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
PTF Operator
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
@@ -7559,7 +7598,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
@@ -7579,7 +7618,7 @@ STAGE PLANS:
serialization.ddl struct part_orc { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment}
serialization.format 1
serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
- totalSize 2599
+ totalSize 2639
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
name: default.part_orc
@@ -7620,12 +7659,14 @@ STAGE PLANS:
Map-reduce partition columns: _col2 (type: string)
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
tag: -1
- value expressions: _col1 (type: string), _col2 (type: string), _col5 (type: int)
+ value expressions: _col5 (type: int)
auto parallelism: true
Reducer 4
Needs Tagging: false
Reduce Operator Tree:
- Extract
+ Select Operator
+ expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col3 (type: int)
+ outputColumnNames: _col1, _col2, _col5
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
PTF Operator
Statistics: Num rows: 26 Data size: 16042 Basic stats: COMPLETE Column stats: NONE
Modified: hive/branches/llap/ql/src/test/results/clientpositive/transform_ppr1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/transform_ppr1.q.out?rev=1654355&r1=1654354&r2=1654355&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/transform_ppr1.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/transform_ppr1.q.out Fri Jan 23 19:59:11 2015
@@ -1,4 +1,6 @@
-PREHOOK: query: EXPLAIN EXTENDED
+PREHOOK: query: -- SORT_QUERY_RESULTS
+
+EXPLAIN EXTENDED
FROM (
FROM srcpart src
SELECT TRANSFORM(src.ds, src.key, src.value)
@@ -7,7 +9,9 @@ FROM (
) tmap
SELECT tmap.tkey, tmap.tvalue WHERE tmap.tkey < 100 AND tmap.ds = '2008-04-08'
PREHOOK: type: QUERY
-POSTHOOK: query: EXPLAIN EXTENDED
+POSTHOOK: query: -- SORT_QUERY_RESULTS
+
+EXPLAIN EXTENDED
FROM (
FROM srcpart src
SELECT TRANSFORM(src.ds, src.key, src.value)
Modified: hive/branches/llap/ql/src/test/results/clientpositive/transform_ppr2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/transform_ppr2.q.out?rev=1654355&r1=1654354&r2=1654355&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/transform_ppr2.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/transform_ppr2.q.out Fri Jan 23 19:59:11 2015
@@ -1,4 +1,6 @@
-PREHOOK: query: EXPLAIN EXTENDED
+PREHOOK: query: -- SORT_QUERY_RESULTS
+
+EXPLAIN EXTENDED
FROM (
FROM srcpart src
SELECT TRANSFORM(src.ds, src.key, src.value)
@@ -8,7 +10,9 @@ FROM (
) tmap
SELECT tmap.tkey, tmap.tvalue WHERE tmap.tkey < 100
PREHOOK: type: QUERY
-POSTHOOK: query: EXPLAIN EXTENDED
+POSTHOOK: query: -- SORT_QUERY_RESULTS
+
+EXPLAIN EXTENDED
FROM (
FROM srcpart src
SELECT TRANSFORM(src.ds, src.key, src.value)
Modified: hive/branches/llap/ql/src/test/results/clientpositive/udf_date_add.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/udf_date_add.q.out?rev=1654355&r1=1654354&r2=1654355&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/udf_date_add.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/udf_date_add.q.out Fri Jan 23 19:59:11 2015
@@ -10,5 +10,5 @@ POSTHOOK: type: DESCFUNCTION
date_add(start_date, num_days) - Returns the date that is num_days after start_date.
start_date is a string in the format 'yyyy-MM-dd HH:mm:ss' or 'yyyy-MM-dd'. num_days is a number. The time part of start_date is ignored.
Example:
- > SELECT date_add('2009-30-07', 1) FROM src LIMIT 1;
- '2009-31-07'
+ > SELECT date_add('2009-07-30', 1) FROM src LIMIT 1;
+ '2009-07-31'
Modified: hive/branches/llap/ql/src/test/results/clientpositive/udf_date_sub.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/udf_date_sub.q.out?rev=1654355&r1=1654354&r2=1654355&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/udf_date_sub.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/udf_date_sub.q.out Fri Jan 23 19:59:11 2015
@@ -10,5 +10,5 @@ POSTHOOK: type: DESCFUNCTION
date_sub(start_date, num_days) - Returns the date that is num_days before start_date.
start_date is a string in the format 'yyyy-MM-dd HH:mm:ss' or 'yyyy-MM-dd'. num_days is a number. The time part of start_date is ignored.
Example:
- > SELECT date_sub('2009-30-07', 1) FROM src LIMIT 1;
- '2009-29-07'
+ > SELECT date_sub('2009-07-30', 1) FROM src LIMIT 1;
+ '2009-07-29'
Modified: hive/branches/llap/ql/src/test/results/clientpositive/udf_datediff.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/udf_datediff.q.out?rev=1654355&r1=1654354&r2=1654355&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/udf_datediff.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/udf_datediff.q.out Fri Jan 23 19:59:11 2015
@@ -10,5 +10,5 @@ POSTHOOK: type: DESCFUNCTION
datediff(date1, date2) - Returns the number of days between date1 and date2
date1 and date2 are strings in the format 'yyyy-MM-dd HH:mm:ss' or 'yyyy-MM-dd'. The time parts are ignored.If date1 is earlier than date2, the result is negative.
Example:
- > SELECT datediff('2009-30-07', '2009-31-07') FROM src LIMIT 1;
+ > SELECT datediff('2009-07-30', '2009-07-31') FROM src LIMIT 1;
1
Modified: hive/branches/llap/ql/src/test/results/clientpositive/udf_day.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/udf_day.q.out?rev=1654355&r1=1654354&r2=1654355&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/udf_day.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/udf_day.q.out Fri Jan 23 19:59:11 2015
@@ -11,5 +11,5 @@ day(date) - Returns the date of the mont
Synonyms: dayofmonth
date is a string in the format of 'yyyy-MM-dd HH:mm:ss' or 'yyyy-MM-dd'.
Example:
- > SELECT day('2009-30-07', 1) FROM src LIMIT 1;
+ > SELECT day('2009-07-30') FROM src LIMIT 1;
30
Modified: hive/branches/llap/ql/src/test/results/clientpositive/udf_dayofmonth.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/udf_dayofmonth.q.out?rev=1654355&r1=1654354&r2=1654355&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/udf_dayofmonth.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/udf_dayofmonth.q.out Fri Jan 23 19:59:11 2015
@@ -11,5 +11,5 @@ dayofmonth(date) - Returns the date of t
Synonyms: day
date is a string in the format of 'yyyy-MM-dd HH:mm:ss' or 'yyyy-MM-dd'.
Example:
- > SELECT dayofmonth('2009-30-07', 1) FROM src LIMIT 1;
+ > SELECT dayofmonth('2009-07-30') FROM src LIMIT 1;
30
Modified: hive/branches/llap/ql/src/test/results/clientpositive/union10.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/union10.q.out?rev=1654355&r1=1654354&r2=1654355&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/union10.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/union10.q.out Fri Jan 23 19:59:11 2015
@@ -1,10 +1,14 @@
-PREHOOK: query: -- union case: all subqueries are a map-reduce jobs, 3 way union, same input for all sub-queries, followed by filesink
+PREHOOK: query: -- SORT_QUERY_RESULTS
+
+-- union case: all subqueries are a map-reduce jobs, 3 way union, same input for all sub-queries, followed by filesink
create table tmptable(key string, value int)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
PREHOOK: Output: default@tmptable
-POSTHOOK: query: -- union case: all subqueries are a map-reduce jobs, 3 way union, same input for all sub-queries, followed by filesink
+POSTHOOK: query: -- SORT_QUERY_RESULTS
+
+-- union case: all subqueries are a map-reduce jobs, 3 way union, same input for all sub-queries, followed by filesink
create table tmptable(key string, value int)
POSTHOOK: type: CREATETABLE
Modified: hive/branches/llap/ql/src/test/results/clientpositive/union18.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/union18.q.out?rev=1654355&r1=1654354&r2=1654355&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/union18.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/union18.q.out Fri Jan 23 19:59:11 2015
@@ -14,7 +14,9 @@ POSTHOOK: query: CREATE TABLE DEST2(key
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
POSTHOOK: Output: default@DEST2
-PREHOOK: query: -- union case:map-reduce sub-queries followed by multi-table insert
+PREHOOK: query: -- SORT_QUERY_RESULTS
+
+-- union case:map-reduce sub-queries followed by multi-table insert
explain
FROM (select 'tst1' as key, cast(count(1) as string) as value from src s1
@@ -23,7 +25,9 @@ FROM (select 'tst1' as key, cast(count(1
INSERT OVERWRITE TABLE DEST1 SELECT unionsrc.key, unionsrc.value
INSERT OVERWRITE TABLE DEST2 SELECT unionsrc.key, unionsrc.value, unionsrc.value
PREHOOK: type: QUERY
-POSTHOOK: query: -- union case:map-reduce sub-queries followed by multi-table insert
+POSTHOOK: query: -- SORT_QUERY_RESULTS
+
+-- union case:map-reduce sub-queries followed by multi-table insert
explain
FROM (select 'tst1' as key, cast(count(1) as string) as value from src s1
Modified: hive/branches/llap/ql/src/test/results/clientpositive/union19.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/union19.q.out?rev=1654355&r1=1654354&r2=1654355&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/union19.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/union19.q.out Fri Jan 23 19:59:11 2015
@@ -1,8 +1,12 @@
-PREHOOK: query: CREATE TABLE DEST1(key STRING, value STRING) STORED AS TEXTFILE
+PREHOOK: query: -- SORT_QUERY_RESULTS
+
+CREATE TABLE DEST1(key STRING, value STRING) STORED AS TEXTFILE
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
PREHOOK: Output: default@DEST1
-POSTHOOK: query: CREATE TABLE DEST1(key STRING, value STRING) STORED AS TEXTFILE
+POSTHOOK: query: -- SORT_QUERY_RESULTS
+
+CREATE TABLE DEST1(key STRING, value STRING) STORED AS TEXTFILE
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
POSTHOOK: Output: default@DEST1
Modified: hive/branches/llap/ql/src/test/results/clientpositive/union27.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/union27.q.out?rev=1654355&r1=1654354&r2=1654355&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/union27.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/union27.q.out Fri Jan 23 19:59:11 2015
@@ -57,7 +57,7 @@ STAGE PLANS:
Reduce Output Operator
key expressions: '97' (type: string)
sort order: +
- Map-reduce partition columns: '' (type: string)
+ Map-reduce partition columns: '97' (type: string)
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
value expressions: _col1 (type: string)
TableScan
@@ -75,7 +75,7 @@ STAGE PLANS:
Reduce Output Operator
key expressions: '97' (type: string)
sort order: +
- Map-reduce partition columns: '' (type: string)
+ Map-reduce partition columns: '97' (type: string)
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
value expressions: _col1 (type: string)
TableScan
@@ -87,7 +87,7 @@ STAGE PLANS:
Reduce Output Operator
key expressions: '97' (type: string)
sort order: +
- Map-reduce partition columns: '' (type: string)
+ Map-reduce partition columns: '97' (type: string)
Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
Reduce Operator Tree:
Join Operator
Modified: hive/branches/llap/ql/src/test/results/clientpositive/union6.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/union6.q.out?rev=1654355&r1=1654354&r2=1654355&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/union6.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/union6.q.out Fri Jan 23 19:59:11 2015
@@ -1,12 +1,14 @@
-PREHOOK: query: -- union case: 1 subquery is a map-reduce job, different inputs for sub-queries, followed by filesink
+PREHOOK: query: -- SORT_QUERY_RESULTS
+-- union case: 1 subquery is a map-reduce job, different inputs for sub-queries, followed by filesink
create table tmptable(key string, value string)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
PREHOOK: Output: default@tmptable
-POSTHOOK: query: -- union case: 1 subquery is a map-reduce job, different inputs for sub-queries, followed by filesink
+POSTHOOK: query: -- SORT_QUERY_RESULTS
+-- union case: 1 subquery is a map-reduce job, different inputs for sub-queries, followed by filesink
create table tmptable(key string, value string)
POSTHOOK: type: CREATETABLE
Modified: hive/branches/llap/ql/src/test/results/clientpositive/union_ppr.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/union_ppr.q.out?rev=1654355&r1=1654354&r2=1654355&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/union_ppr.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/union_ppr.q.out Fri Jan 23 19:59:11 2015
@@ -1,4 +1,6 @@
-PREHOOK: query: EXPLAIN EXTENDED
+PREHOOK: query: -- SORT_QUERY_RESULTS
+
+EXPLAIN EXTENDED
SELECT * FROM (
SELECT X.* FROM SRCPART X WHERE X.key < 100
UNION ALL
@@ -7,7 +9,9 @@ SELECT * FROM (
WHERE A.ds = '2008-04-08'
SORT BY A.key, A.value, A.ds, A.hr
PREHOOK: type: QUERY
-POSTHOOK: query: EXPLAIN EXTENDED
+POSTHOOK: query: -- SORT_QUERY_RESULTS
+
+EXPLAIN EXTENDED
SELECT * FROM (
SELECT X.* FROM SRCPART X WHERE X.key < 100
UNION ALL
Modified: hive/branches/llap/ql/src/test/results/clientpositive/vector_left_outer_join.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/vector_left_outer_join.q.out?rev=1654355&r1=1654354&r2=1654355&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/vector_left_outer_join.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/vector_left_outer_join.q.out Fri Jan 23 19:59:11 2015
@@ -25,14 +25,14 @@ STAGE PLANS:
Stage: Stage-8
Map Reduce Local Work
Alias -> Map Local Tables:
- $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_1:c
+ $hdt$_0:$hdt$_1:c
Fetch Operator
limit: -1
- $hdt$_0:$hdt$_0:$hdt$_1:c
+ $hdt$_0:$hdt$_2:c
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_1:c
+ $hdt$_0:$hdt$_1:c
TableScan
alias: c
Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
@@ -44,7 +44,7 @@ STAGE PLANS:
keys:
0 _col1 (type: int)
1 _col0 (type: int)
- $hdt$_0:$hdt$_0:$hdt$_1:c
+ $hdt$_0:$hdt$_2:c
TableScan
alias: c
Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
Modified: hive/branches/llap/ql/src/test/results/clientpositive/vector_mapjoin_reduce.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/vector_mapjoin_reduce.q.out?rev=1654355&r1=1654354&r2=1654355&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/vector_mapjoin_reduce.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/vector_mapjoin_reduce.q.out Fri Jan 23 19:59:11 2015
@@ -475,7 +475,7 @@ STAGE PLANS:
Statistics: Num rows: 13 Data size: 1559 Basic stats: COMPLETE Column stats: NONE
HashTable Sink Operator
keys:
- 0 _col0 (type: int), 1 (type: int)
+ 0 _col0 (type: int), _col3 (type: int)
1 _col0 (type: int), _col1 (type: int)
Stage: Stage-8
@@ -488,14 +488,14 @@ STAGE PLANS:
predicate: (((l_linenumber = 1) and l_orderkey is not null) and l_partkey is not null) (type: boolean)
Statistics: Num rows: 13 Data size: 1559 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: l_orderkey (type: int), l_partkey (type: int), l_suppkey (type: int)
- outputColumnNames: _col0, _col1, _col2
+ expressions: l_orderkey (type: int), l_partkey (type: int), l_suppkey (type: int), 1 (type: int)
+ outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 13 Data size: 1559 Basic stats: COMPLETE Column stats: NONE
Map Join Operator
condition map:
Left Semi Join 0 to 1
keys:
- 0 _col0 (type: int), 1 (type: int)
+ 0 _col0 (type: int), _col3 (type: int)
1 _col0 (type: int), _col1 (type: int)
outputColumnNames: _col1, _col2
Statistics: Num rows: 14 Data size: 1714 Basic stats: COMPLETE Column stats: NONE
Modified: hive/branches/llap/ql/src/test/results/clientpositive/vector_string_concat.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/vector_string_concat.q.out?rev=1654355&r1=1654354&r2=1654355&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/vector_string_concat.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/vector_string_concat.q.out Fri Jan 23 19:59:11 2015
@@ -285,17 +285,20 @@ PREHOOK: query: EXPLAIN
SELECT CONCAT(CONCAT(CONCAT('Quarter ',CAST(CAST((MONTH(dt) - 1) / 3 + 1 AS INT) AS STRING)),'-'),CAST(YEAR(dt) AS STRING)) AS `field`
FROM vectortab2korc
GROUP BY CONCAT(CONCAT(CONCAT('Quarter ',CAST(CAST((MONTH(dt) - 1) / 3 + 1 AS INT) AS STRING)),'-'),CAST(YEAR(dt) AS STRING))
+ ORDER BY `field`
LIMIT 50
PREHOOK: type: QUERY
POSTHOOK: query: EXPLAIN
SELECT CONCAT(CONCAT(CONCAT('Quarter ',CAST(CAST((MONTH(dt) - 1) / 3 + 1 AS INT) AS STRING)),'-'),CAST(YEAR(dt) AS STRING)) AS `field`
FROM vectortab2korc
GROUP BY CONCAT(CONCAT(CONCAT('Quarter ',CAST(CAST((MONTH(dt) - 1) / 3 + 1 AS INT) AS STRING)),'-'),CAST(YEAR(dt) AS STRING))
+ ORDER BY `field`
LIMIT 50
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
Stage-1 is a root stage
- Stage-0 depends on stages: Stage-1
+ Stage-2 depends on stages: Stage-1
+ Stage-0 depends on stages: Stage-2
STAGE PLANS:
Stage: Stage-1
@@ -325,20 +328,36 @@ STAGE PLANS:
mode: mergepartial
outputColumnNames: _col0
Statistics: Num rows: 1000 Data size: 459356 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string)
- outputColumnNames: _col0
- Statistics: Num rows: 1000 Data size: 459356 Basic stats: COMPLETE Column stats: NONE
- Limit
- Number of rows: 50
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
+
+ Stage: Stage-2
+ Map Reduce
+ Map Operator Tree:
+ TableScan
+ Reduce Output Operator
+ key expressions: _col0 (type: string)
+ sort order: +
+ Statistics: Num rows: 1000 Data size: 459356 Basic stats: COMPLETE Column stats: NONE
+ Reduce Operator Tree:
+ Select Operator
+ expressions: KEY.reducesinkkey0 (type: string)
+ outputColumnNames: _col0
+ Statistics: Num rows: 1000 Data size: 459356 Basic stats: COMPLETE Column stats: NONE
+ Limit
+ Number of rows: 50
+ Statistics: Num rows: 50 Data size: 22950 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
Statistics: Num rows: 50 Data size: 22950 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 50 Data size: 22950 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Stage: Stage-0
Fetch Operator
@@ -349,6 +368,7 @@ STAGE PLANS:
PREHOOK: query: SELECT CONCAT(CONCAT(CONCAT('Quarter ',CAST(CAST((MONTH(dt) - 1) / 3 + 1 AS INT) AS STRING)),'-'),CAST(YEAR(dt) AS STRING)) AS `field`
FROM vectortab2korc
GROUP BY CONCAT(CONCAT(CONCAT('Quarter ',CAST(CAST((MONTH(dt) - 1) / 3 + 1 AS INT) AS STRING)),'-'),CAST(YEAR(dt) AS STRING))
+ ORDER BY `field`
LIMIT 50
PREHOOK: type: QUERY
PREHOOK: Input: default@vectortab2korc
@@ -356,6 +376,7 @@ PREHOOK: Input: default@vectortab2korc
POSTHOOK: query: SELECT CONCAT(CONCAT(CONCAT('Quarter ',CAST(CAST((MONTH(dt) - 1) / 3 + 1 AS INT) AS STRING)),'-'),CAST(YEAR(dt) AS STRING)) AS `field`
FROM vectortab2korc
GROUP BY CONCAT(CONCAT(CONCAT('Quarter ',CAST(CAST((MONTH(dt) - 1) / 3 + 1 AS INT) AS STRING)),'-'),CAST(YEAR(dt) AS STRING))
+ ORDER BY `field`
LIMIT 50
POSTHOOK: type: QUERY
POSTHOOK: Input: default@vectortab2korc
Modified: hive/branches/llap/ql/src/test/results/clientpositive/vectorized_mapjoin.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/vectorized_mapjoin.q.out?rev=1654355&r1=1654354&r2=1654355&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/vectorized_mapjoin.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/vectorized_mapjoin.q.out Fri Jan 23 19:59:11 2015
@@ -15,11 +15,11 @@ STAGE PLANS:
Stage: Stage-5
Map Reduce Local Work
Alias -> Map Local Tables:
- $hdt$_0:$hdt$_0:$hdt$_0:t1
+ $hdt$_0:$hdt$_0:t1
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- $hdt$_0:$hdt$_0:$hdt$_0:t1
+ $hdt$_0:$hdt$_0:t1
TableScan
alias: t1
Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
Modified: hive/branches/llap/ql/src/test/results/clientpositive/vectorized_nested_mapjoin.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/vectorized_nested_mapjoin.q.out?rev=1654355&r1=1654354&r2=1654355&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/vectorized_nested_mapjoin.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/vectorized_nested_mapjoin.q.out Fri Jan 23 19:59:11 2015
@@ -11,14 +11,14 @@ STAGE PLANS:
Stage: Stage-8
Map Reduce Local Work
Alias -> Map Local Tables:
- $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:v1
+ $hdt$_0:$hdt$_0:$hdt$_0:v1
Fetch Operator
limit: -1
- $hdt$_0:$hdt$_0:$hdt$_1:v1
+ $hdt$_0:$hdt$_1:v1
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- $hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:$hdt$_0:v1
+ $hdt$_0:$hdt$_0:$hdt$_0:v1
TableScan
alias: v1
Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
@@ -33,7 +33,7 @@ STAGE PLANS:
keys:
0 _col0 (type: tinyint)
1 _col0 (type: tinyint)
- $hdt$_0:$hdt$_0:$hdt$_1:v1
+ $hdt$_0:$hdt$_1:v1
TableScan
alias: v1
Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE