You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by xu...@apache.org on 2014/11/27 14:32:49 UTC
svn commit: r1642149 - in /hive/branches/spark: ./ hbase-handler/
ql/src/test/queries/clientpositive/ ql/src/test/results/clientpositive/
ql/src/test/results/clientpositive/tez/
Author: xuefu
Date: Thu Nov 27 13:32:48 2014
New Revision: 1642149
URL: http://svn.apache.org/r1642149
Log:
HIVE-8976: Make nine additional tests deterministic (Brock via Xuefu)
merged from trunk, r1642148
Modified:
hive/branches/spark/ (props changed)
hive/branches/spark/hbase-handler/pom.xml (props changed)
hive/branches/spark/ql/src/test/queries/clientpositive/auto_join_without_localtask.q
hive/branches/spark/ql/src/test/queries/clientpositive/count.q
hive/branches/spark/ql/src/test/queries/clientpositive/limit_pushdown.q
hive/branches/spark/ql/src/test/queries/clientpositive/mapreduce2.q
hive/branches/spark/ql/src/test/queries/clientpositive/multi_insert_gby3.q
hive/branches/spark/ql/src/test/queries/clientpositive/multi_join_union.q
hive/branches/spark/ql/src/test/queries/clientpositive/ppd_outer_join3.q
hive/branches/spark/ql/src/test/queries/clientpositive/ptf_decimal.q
hive/branches/spark/ql/src/test/queries/clientpositive/ptf_general_queries.q
hive/branches/spark/ql/src/test/results/clientpositive/auto_join_without_localtask.q.out
hive/branches/spark/ql/src/test/results/clientpositive/count.q.out
hive/branches/spark/ql/src/test/results/clientpositive/limit_pushdown.q.out
hive/branches/spark/ql/src/test/results/clientpositive/mapreduce2.q.out
hive/branches/spark/ql/src/test/results/clientpositive/multi_insert_gby3.q.out
hive/branches/spark/ql/src/test/results/clientpositive/multi_join_union.q.out
hive/branches/spark/ql/src/test/results/clientpositive/ppd_outer_join3.q.out
hive/branches/spark/ql/src/test/results/clientpositive/ptf_decimal.q.out
hive/branches/spark/ql/src/test/results/clientpositive/ptf_general_queries.q.out
hive/branches/spark/ql/src/test/results/clientpositive/tez/count.q.out
hive/branches/spark/ql/src/test/results/clientpositive/tez/limit_pushdown.q.out
hive/branches/spark/ql/src/test/results/clientpositive/tez/mapreduce2.q.out
Propchange: hive/branches/spark/
------------------------------------------------------------------------------
Merged /hive/trunk:r1642148
Propchange: hive/branches/spark/hbase-handler/pom.xml
------------------------------------------------------------------------------
Merged /hive/trunk/hbase-handler/pom.xml:r1642148
Modified: hive/branches/spark/ql/src/test/queries/clientpositive/auto_join_without_localtask.q
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/queries/clientpositive/auto_join_without_localtask.q?rev=1642149&r1=1642148&r2=1642149&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/queries/clientpositive/auto_join_without_localtask.q (original)
+++ hive/branches/spark/ql/src/test/queries/clientpositive/auto_join_without_localtask.q Thu Nov 27 13:32:48 2014
@@ -5,14 +5,14 @@ set hive.auto.convert.join.use.nonstaged
set hive.auto.convert.join.noconditionaltask.size=100;
explain
-select a.* from src a join src b on a.key=b.key limit 40;
+select a.* from src a join src b on a.key=b.key order by key, value limit 40;
-select a.* from src a join src b on a.key=b.key limit 40;
+select a.* from src a join src b on a.key=b.key order by key, value limit 40;
explain
-select a.* from src a join src b on a.key=b.key join src c on a.value=c.value limit 40;
+select a.* from src a join src b on a.key=b.key join src c on a.value=c.value order by a.key, a.value limit 40;
-select a.* from src a join src b on a.key=b.key join src c on a.value=c.value limit 40;
+select a.* from src a join src b on a.key=b.key join src c on a.value=c.value order by a.key, a.value limit 40;
set hive.auto.convert.join.noconditionaltask.size=100;
Modified: hive/branches/spark/ql/src/test/queries/clientpositive/count.q
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/queries/clientpositive/count.q?rev=1642149&r1=1642148&r2=1642149&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/queries/clientpositive/count.q (original)
+++ hive/branches/spark/ql/src/test/queries/clientpositive/count.q Thu Nov 27 13:32:48 2014
@@ -1,3 +1,4 @@
+-- SORT_QUERY_RESULTS
create table abcd (a int, b int, c int, d int);
LOAD DATA LOCAL INPATH '../../data/files/in4.txt' INTO TABLE abcd;
Modified: hive/branches/spark/ql/src/test/queries/clientpositive/limit_pushdown.q
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/queries/clientpositive/limit_pushdown.q?rev=1642149&r1=1642148&r2=1642149&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/queries/clientpositive/limit_pushdown.q (original)
+++ hive/branches/spark/ql/src/test/queries/clientpositive/limit_pushdown.q Thu Nov 27 13:32:48 2014
@@ -14,8 +14,8 @@ select key,value from src order by key d
select key,value from src order by key desc limit 20;
explain
-select value, sum(key + 1) as sum from src group by value limit 20;
-select value, sum(key + 1) as sum from src group by value limit 20;
+select value, sum(key + 1) as sum from src group by value order by value limit 20;
+select value, sum(key + 1) as sum from src group by value order by value limit 20;
-- deduped RS
explain
@@ -28,13 +28,13 @@ select distinct(cdouble) from alltypesor
select distinct(cdouble) from alltypesorc limit 20;
explain
-select ctinyint, count(distinct(cdouble)) from alltypesorc group by ctinyint limit 20;
-select ctinyint, count(distinct(cdouble)) from alltypesorc group by ctinyint limit 20;
+select ctinyint, count(distinct(cdouble)) from alltypesorc group by ctinyint order by ctinyint limit 20;
+select ctinyint, count(distinct(cdouble)) from alltypesorc group by ctinyint order by ctinyint limit 20;
-- multi distinct
explain
-select ctinyint, count(distinct(cstring1)), count(distinct(cstring2)) from alltypesorc group by ctinyint limit 20;
-select ctinyint, count(distinct(cstring1)), count(distinct(cstring2)) from alltypesorc group by ctinyint limit 20;
+select ctinyint, count(distinct(cstring1)), count(distinct(cstring2)) from alltypesorc group by ctinyint order by ctinyint limit 20;
+select ctinyint, count(distinct(cstring1)), count(distinct(cstring2)) from alltypesorc group by ctinyint order by ctinyint limit 20;
-- limit zero
explain
@@ -57,8 +57,8 @@ on subq.key=subq2.key limit 4;
set hive.map.aggr=false;
-- map aggregation disabled
explain
-select value, sum(key) as sum from src group by value limit 20;
-select value, sum(key) as sum from src group by value limit 20;
+select value, sum(key) as sum from src group by value order by value limit 20;
+select value, sum(key) as sum from src group by value order by value limit 20;
set hive.limit.pushdown.memory.usage=0.00002f;
Modified: hive/branches/spark/ql/src/test/queries/clientpositive/mapreduce2.q
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/queries/clientpositive/mapreduce2.q?rev=1642149&r1=1642148&r2=1642149&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/queries/clientpositive/mapreduce2.q (original)
+++ hive/branches/spark/ql/src/test/queries/clientpositive/mapreduce2.q Thu Nov 27 13:32:48 2014
@@ -14,4 +14,4 @@ MAP src.key, CAST(src.key / 10 AS INT),
USING 'cat' AS (tkey, ten, one, tvalue)
DISTRIBUTE BY tvalue, tkey;
-SELECT * FROM (SELECT dest1.* FROM dest1 DISTRIBUTE BY key SORT BY key, ten, one, value) T;
+SELECT * FROM (SELECT dest1.* FROM dest1 DISTRIBUTE BY key SORT BY key, ten, one, value) T ORDER BY key;
Modified: hive/branches/spark/ql/src/test/queries/clientpositive/multi_insert_gby3.q
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/queries/clientpositive/multi_insert_gby3.q?rev=1642149&r1=1642148&r2=1642149&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/queries/clientpositive/multi_insert_gby3.q (original)
+++ hive/branches/spark/ql/src/test/queries/clientpositive/multi_insert_gby3.q Thu Nov 27 13:32:48 2014
@@ -1,3 +1,4 @@
+-- SORT_QUERY_RESULTS
create table e1 (key string, keyD double);
create table e2 (key string, keyD double, value string);
create table e3 (key string, keyD double);
Modified: hive/branches/spark/ql/src/test/queries/clientpositive/multi_join_union.q
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/queries/clientpositive/multi_join_union.q?rev=1642149&r1=1642148&r2=1642149&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/queries/clientpositive/multi_join_union.q (original)
+++ hive/branches/spark/ql/src/test/queries/clientpositive/multi_join_union.q Thu Nov 27 13:32:48 2014
@@ -1,6 +1,7 @@
-
set hive.auto.convert.join=true;
+-- SORT_QUERY_RESULTS
+
CREATE TABLE src11 as SELECT * FROM src;
CREATE TABLE src12 as SELECT * FROM src;
CREATE TABLE src13 as SELECT * FROM src;
@@ -10,4 +11,4 @@ CREATE TABLE src14 as SELECT * FROM src;
EXPLAIN SELECT * FROM
src11 a JOIN
src12 b ON (a.key = b.key) JOIN
-(SELECT * FROM (SELECT * FROM src13 UNION ALL SELECT * FROM src14)a )c ON c.value = b.value;
\ No newline at end of file
+(SELECT * FROM (SELECT * FROM src13 UNION ALL SELECT * FROM src14)a )c ON c.value = b.value;
Modified: hive/branches/spark/ql/src/test/queries/clientpositive/ppd_outer_join3.q
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/queries/clientpositive/ppd_outer_join3.q?rev=1642149&r1=1642148&r2=1642149&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/queries/clientpositive/ppd_outer_join3.q (original)
+++ hive/branches/spark/ql/src/test/queries/clientpositive/ppd_outer_join3.q Thu Nov 27 13:32:48 2014
@@ -1,6 +1,8 @@
set hive.optimize.ppd=true;
set hive.ppd.remove.duplicatefilters=false;
+-- SORT_QUERY_RESULTS
+
EXPLAIN
FROM
src a
Modified: hive/branches/spark/ql/src/test/queries/clientpositive/ptf_decimal.q
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/queries/clientpositive/ptf_decimal.q?rev=1642149&r1=1642148&r2=1642149&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/queries/clientpositive/ptf_decimal.q (original)
+++ hive/branches/spark/ql/src/test/queries/clientpositive/ptf_decimal.q Thu Nov 27 13:32:48 2014
@@ -1,3 +1,4 @@
+-- SORT_QUERY_RESULTS
-- 1. aggregate functions with decimal type
Modified: hive/branches/spark/ql/src/test/queries/clientpositive/ptf_general_queries.q
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/queries/clientpositive/ptf_general_queries.q?rev=1642149&r1=1642148&r2=1642149&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/queries/clientpositive/ptf_general_queries.q (original)
+++ hive/branches/spark/ql/src/test/queries/clientpositive/ptf_general_queries.q Thu Nov 27 13:32:48 2014
@@ -1,3 +1,5 @@
+-- SORT_QUERY_RESULTS
+
-- 1. testNoPTFNoWindowing
select p_mfgr, p_name, p_size
from part
Modified: hive/branches/spark/ql/src/test/results/clientpositive/auto_join_without_localtask.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/auto_join_without_localtask.q.out?rev=1642149&r1=1642148&r2=1642149&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/auto_join_without_localtask.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/auto_join_without_localtask.q.out Thu Nov 27 13:32:48 2014
@@ -1,23 +1,24 @@
PREHOOK: query: explain
-select a.* from src a join src b on a.key=b.key limit 40
+select a.* from src a join src b on a.key=b.key order by key, value limit 40
PREHOOK: type: QUERY
POSTHOOK: query: explain
-select a.* from src a join src b on a.key=b.key limit 40
+select a.* from src a join src b on a.key=b.key order by key, value limit 40
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-5 is a root stage , consists of Stage-6, Stage-7, Stage-1
- Stage-6 has a backup stage: Stage-1
- Stage-3 depends on stages: Stage-6
+ Stage-6 is a root stage , consists of Stage-7, Stage-8, Stage-1
Stage-7 has a backup stage: Stage-1
Stage-4 depends on stages: Stage-7
+ Stage-2 depends on stages: Stage-1, Stage-4, Stage-5
+ Stage-8 has a backup stage: Stage-1
+ Stage-5 depends on stages: Stage-8
Stage-1
- Stage-0 depends on stages: Stage-3, Stage-4, Stage-1
+ Stage-0 depends on stages: Stage-2
STAGE PLANS:
- Stage: Stage-5
+ Stage: Stage-6
Conditional Operator
- Stage: Stage-6
+ Stage: Stage-7
Map Reduce Local Work
Alias -> Map Local Tables:
b
@@ -37,7 +38,7 @@ STAGE PLANS:
0 key (type: string)
1 key (type: string)
- Stage: Stage-3
+ Stage: Stage-4
Map Reduce
Map Operator Tree:
TableScan
@@ -54,21 +55,40 @@ STAGE PLANS:
0 key (type: string)
1 key (type: string)
outputColumnNames: _col0, _col1
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- Limit
- Number of rows: 40
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
Local Work:
Map Reduce Local Work
- Stage: Stage-7
+ Stage: Stage-2
+ Map Reduce
+ Map Operator Tree:
+ TableScan
+ Reduce Output Operator
+ key expressions: _col0 (type: string), _col1 (type: string)
+ sort order: ++
+ Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
+ Reduce Operator Tree:
+ Select Operator
+ expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
+ Limit
+ Number of rows: 40
+ Statistics: Num rows: 40 Data size: 400 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 40 Data size: 400 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+ Stage: Stage-8
Map Reduce Local Work
Alias -> Map Local Tables:
a
@@ -88,7 +108,7 @@ STAGE PLANS:
0 key (type: string)
1 key (type: string)
- Stage: Stage-4
+ Stage: Stage-5
Map Reduce
Map Operator Tree:
TableScan
@@ -105,17 +125,12 @@ STAGE PLANS:
0 key (type: string)
1 key (type: string)
outputColumnNames: _col0, _col1
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- Limit
- Number of rows: 40
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
Local Work:
Map Reduce Local Work
@@ -154,20 +169,12 @@ STAGE PLANS:
1
outputColumnNames: _col0, _col1
Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
- Limit
- Number of rows: 40
- Statistics: Num rows: 40 Data size: 400 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 40 Data size: 400 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
Stage: Stage-0
Fetch Operator
@@ -175,82 +182,84 @@ STAGE PLANS:
Processor Tree:
ListSink
-PREHOOK: query: select a.* from src a join src b on a.key=b.key limit 40
+PREHOOK: query: select a.* from src a join src b on a.key=b.key order by key, value limit 40
PREHOOK: type: QUERY
PREHOOK: Input: default@src
#### A masked pattern was here ####
-POSTHOOK: query: select a.* from src a join src b on a.key=b.key limit 40
+POSTHOOK: query: select a.* from src a join src b on a.key=b.key order by key, value limit 40
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
#### A masked pattern was here ####
-RUN: Stage-5:CONDITIONAL
+RUN: Stage-6:CONDITIONAL
RUN: Stage-7:MAPREDLOCAL
RUN: Stage-4:MAPRED
-238 val_238
-238 val_238
-86 val_86
-311 val_311
-311 val_311
-311 val_311
-27 val_27
-165 val_165
-165 val_165
-409 val_409
-409 val_409
-409 val_409
-255 val_255
-255 val_255
-278 val_278
-278 val_278
-98 val_98
-98 val_98
-484 val_484
-265 val_265
-265 val_265
-193 val_193
-193 val_193
-193 val_193
-401 val_401
-401 val_401
-401 val_401
-401 val_401
-401 val_401
-150 val_150
-273 val_273
-273 val_273
-273 val_273
-224 val_224
-224 val_224
-369 val_369
-369 val_369
-369 val_369
-66 val_66
-128 val_128
+RUN: Stage-2:MAPRED
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+10 val_10
+100 val_100
+100 val_100
+100 val_100
+100 val_100
+103 val_103
+103 val_103
+103 val_103
+103 val_103
+104 val_104
+104 val_104
+104 val_104
+104 val_104
+105 val_105
+11 val_11
+111 val_111
+113 val_113
+113 val_113
+113 val_113
+113 val_113
+114 val_114
+116 val_116
+118 val_118
+118 val_118
+118 val_118
+118 val_118
+119 val_119
+119 val_119
+119 val_119
+119 val_119
+119 val_119
PREHOOK: query: explain
-select a.* from src a join src b on a.key=b.key join src c on a.value=c.value limit 40
+select a.* from src a join src b on a.key=b.key join src c on a.value=c.value order by a.key, a.value limit 40
PREHOOK: type: QUERY
POSTHOOK: query: explain
-select a.* from src a join src b on a.key=b.key join src c on a.value=c.value limit 40
+select a.* from src a join src b on a.key=b.key join src c on a.value=c.value order by a.key, a.value limit 40
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-10 is a root stage , consists of Stage-13, Stage-14, Stage-1
- Stage-13 has a backup stage: Stage-1
- Stage-8 depends on stages: Stage-13
- Stage-7 depends on stages: Stage-1, Stage-8, Stage-9 , consists of Stage-11, Stage-6, Stage-2
- Stage-11 has a backup stage: Stage-2
- Stage-5 depends on stages: Stage-11
- Stage-6 has a backup stage: Stage-2
- Stage-2
+ Stage-11 is a root stage , consists of Stage-14, Stage-15, Stage-1
Stage-14 has a backup stage: Stage-1
Stage-9 depends on stages: Stage-14
+ Stage-8 depends on stages: Stage-1, Stage-9, Stage-10 , consists of Stage-12, Stage-7, Stage-2
+ Stage-12 has a backup stage: Stage-2
+ Stage-6 depends on stages: Stage-12
+ Stage-3 depends on stages: Stage-2, Stage-6, Stage-7
+ Stage-7 has a backup stage: Stage-2
+ Stage-2
+ Stage-15 has a backup stage: Stage-1
+ Stage-10 depends on stages: Stage-15
Stage-1
- Stage-0 depends on stages: Stage-5, Stage-6, Stage-2
+ Stage-0 depends on stages: Stage-3
STAGE PLANS:
- Stage: Stage-10
+ Stage: Stage-11
Conditional Operator
- Stage: Stage-13
+ Stage: Stage-14
Map Reduce Local Work
Alias -> Map Local Tables:
b
@@ -270,7 +279,7 @@ STAGE PLANS:
0 key (type: string)
1 key (type: string)
- Stage: Stage-8
+ Stage: Stage-9
Map Reduce
Map Operator Tree:
TableScan
@@ -296,10 +305,10 @@ STAGE PLANS:
Local Work:
Map Reduce Local Work
- Stage: Stage-7
+ Stage: Stage-8
Conditional Operator
- Stage: Stage-11
+ Stage: Stage-12
Map Reduce Local Work
Alias -> Map Local Tables:
c
@@ -319,7 +328,7 @@ STAGE PLANS:
0 _col1 (type: string)
1 value (type: string)
- Stage: Stage-5
+ Stage: Stage-6
Map Reduce
Map Operator Tree:
TableScan
@@ -333,21 +342,40 @@ STAGE PLANS:
0 _col1 (type: string)
1 value (type: string)
outputColumnNames: _col0, _col1
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- Limit
- Number of rows: 40
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
Local Work:
Map Reduce Local Work
- Stage: Stage-6
+ Stage: Stage-3
+ Map Reduce
+ Map Operator Tree:
+ TableScan
+ Reduce Output Operator
+ key expressions: _col0 (type: string), _col1 (type: string)
+ sort order: ++
+ Statistics: Num rows: 302 Data size: 3213 Basic stats: COMPLETE Column stats: NONE
+ Reduce Operator Tree:
+ Select Operator
+ expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 302 Data size: 3213 Basic stats: COMPLETE Column stats: NONE
+ Limit
+ Number of rows: 40
+ Statistics: Num rows: 40 Data size: 400 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 40 Data size: 400 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+ Stage: Stage-7
Map Reduce
Map Operator Tree:
TableScan
@@ -364,17 +392,12 @@ STAGE PLANS:
0 _col1 (type: string)
1 value (type: string)
outputColumnNames: _col0, _col1
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- Limit
- Number of rows: 40
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
Local Work:
Map Reduce Local Work
Alias -> Map Local Tables:
@@ -415,22 +438,14 @@ STAGE PLANS:
1
outputColumnNames: _col0, _col1
Statistics: Num rows: 302 Data size: 3213 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 302 Data size: 3213 Basic stats: COMPLETE Column stats: NONE
- Limit
- Number of rows: 40
- Statistics: Num rows: 40 Data size: 400 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 40 Data size: 400 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
- Stage: Stage-14
+ Stage: Stage-15
Map Reduce Local Work
Alias -> Map Local Tables:
a
@@ -450,7 +465,7 @@ STAGE PLANS:
0 key (type: string)
1 key (type: string)
- Stage: Stage-9
+ Stage: Stage-10
Map Reduce
Map Operator Tree:
TableScan
@@ -524,60 +539,61 @@ STAGE PLANS:
Processor Tree:
ListSink
-PREHOOK: query: select a.* from src a join src b on a.key=b.key join src c on a.value=c.value limit 40
+PREHOOK: query: select a.* from src a join src b on a.key=b.key join src c on a.value=c.value order by a.key, a.value limit 40
PREHOOK: type: QUERY
PREHOOK: Input: default@src
#### A masked pattern was here ####
-POSTHOOK: query: select a.* from src a join src b on a.key=b.key join src c on a.value=c.value limit 40
+POSTHOOK: query: select a.* from src a join src b on a.key=b.key join src c on a.value=c.value order by a.key, a.value limit 40
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
#### A masked pattern was here ####
-RUN: Stage-10:CONDITIONAL
+RUN: Stage-11:CONDITIONAL
RUN: Stage-14:MAPREDLOCAL
RUN: Stage-9:MAPRED
-RUN: Stage-7:CONDITIONAL
-RUN: Stage-11:MAPREDLOCAL
-RUN: Stage-5:MAPRED
-238 val_238
-238 val_238
-238 val_238
-238 val_238
-86 val_86
-311 val_311
-311 val_311
-311 val_311
-311 val_311
-311 val_311
-311 val_311
-311 val_311
-311 val_311
-311 val_311
-27 val_27
-165 val_165
-165 val_165
-165 val_165
-165 val_165
-409 val_409
-409 val_409
-409 val_409
-409 val_409
-409 val_409
-409 val_409
-409 val_409
-409 val_409
-409 val_409
-255 val_255
-255 val_255
-255 val_255
-255 val_255
-278 val_278
-278 val_278
-278 val_278
-278 val_278
-98 val_98
-98 val_98
-98 val_98
-98 val_98
+RUN: Stage-8:CONDITIONAL
+RUN: Stage-12:MAPREDLOCAL
+RUN: Stage-6:MAPRED
+RUN: Stage-3:MAPRED
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+0 val_0
+10 val_10
+100 val_100
+100 val_100
+100 val_100
+100 val_100
+100 val_100
+100 val_100
+100 val_100
+100 val_100
+103 val_103
+103 val_103
+103 val_103
+103 val_103
PREHOOK: query: explain
select a.* from src a join src b on a.key=b.key join src c on a.value=c.value where a.key>100 limit 40
PREHOOK: type: QUERY
Modified: hive/branches/spark/ql/src/test/results/clientpositive/count.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/count.q.out?rev=1642149&r1=1642148&r2=1642149&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/count.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/count.q.out Thu Nov 27 13:32:48 2014
@@ -1,8 +1,10 @@
-PREHOOK: query: create table abcd (a int, b int, c int, d int)
+PREHOOK: query: -- SORT_QUERY_RESULTS
+create table abcd (a int, b int, c int, d int)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
PREHOOK: Output: default@abcd
-POSTHOOK: query: create table abcd (a int, b int, c int, d int)
+POSTHOOK: query: -- SORT_QUERY_RESULTS
+create table abcd (a int, b int, c int, d int)
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
POSTHOOK: Output: default@abcd
@@ -22,13 +24,13 @@ POSTHOOK: query: select * from abcd
POSTHOOK: type: QUERY
POSTHOOK: Input: default@abcd
#### A masked pattern was here ####
-NULL 35 23 6
+10 100 45 4
+10 100 NULL 5
10 1000 50 1
100 100 10 3
-12 NULL 80 2
-10 100 NULL 5
-10 100 45 4
12 100 75 7
+12 NULL 80 2
+NULL 35 23 6
PREHOOK: query: explain select a, count(distinct b), count(distinct c), sum(d) from abcd group by a
PREHOOK: type: QUERY
POSTHOOK: query: explain select a, count(distinct b), count(distinct c), sum(d) from abcd group by a
@@ -93,10 +95,10 @@ POSTHOOK: query: select a, count(distinc
POSTHOOK: type: QUERY
POSTHOOK: Input: default@abcd
#### A masked pattern was here ####
-NULL 1 1 6
10 2 2 10
-12 1 2 9
100 1 1 3
+12 1 2 9
+NULL 1 1 6
PREHOOK: query: explain select count(1), count(*), count(a), count(b), count(c), count(d), count(distinct a), count(distinct b), count(distinct c), count(distinct d), count(distinct a,b), count(distinct b,c), count(distinct c,d), count(distinct a,d), count(distinct a,c), count(distinct b,d), count(distinct a,b,c), count(distinct b,c,d), count(distinct a,c,d), count(distinct a,b,d), count(distinct a,b,c,d) from abcd
PREHOOK: type: QUERY
POSTHOOK: query: explain select count(1), count(*), count(a), count(b), count(c), count(d), count(distinct a), count(distinct b), count(distinct c), count(distinct d), count(distinct a,b), count(distinct b,c), count(distinct c,d), count(distinct a,d), count(distinct a,c), count(distinct b,d), count(distinct a,b,c), count(distinct b,c,d), count(distinct a,c,d), count(distinct a,b,d), count(distinct a,b,c,d) from abcd
@@ -218,10 +220,10 @@ POSTHOOK: query: select a, count(distinc
POSTHOOK: type: QUERY
POSTHOOK: Input: default@abcd
#### A masked pattern was here ####
-NULL 1 1 6
10 2 2 10
-12 1 2 9
100 1 1 3
+12 1 2 9
+NULL 1 1 6
PREHOOK: query: explain select count(1), count(*), count(a), count(b), count(c), count(d), count(distinct a), count(distinct b), count(distinct c), count(distinct d), count(distinct a,b), count(distinct b,c), count(distinct c,d), count(distinct a,d), count(distinct a,c), count(distinct b,d), count(distinct a,b,c), count(distinct b,c,d), count(distinct a,c,d), count(distinct a,b,d), count(distinct a,b,c,d) from abcd
PREHOOK: type: QUERY
POSTHOOK: query: explain select count(1), count(*), count(a), count(b), count(c), count(d), count(distinct a), count(distinct b), count(distinct c), count(distinct d), count(distinct a,b), count(distinct b,c), count(distinct c,d), count(distinct a,d), count(distinct a,c), count(distinct b,d), count(distinct a,b,c), count(distinct b,c,d), count(distinct a,c,d), count(distinct a,b,d), count(distinct a,b,c,d) from abcd
Modified: hive/branches/spark/ql/src/test/results/clientpositive/limit_pushdown.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/limit_pushdown.q.out?rev=1642149&r1=1642148&r2=1642149&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/limit_pushdown.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/limit_pushdown.q.out Thu Nov 27 13:32:48 2014
@@ -161,10 +161,10 @@ POSTHOOK: Input: default@src
98 val_98
98 val_98
PREHOOK: query: explain
-select value, sum(key + 1) as sum from src group by value limit 20
+select value, sum(key + 1) as sum from src group by value order by value limit 20
PREHOOK: type: QUERY
POSTHOOK: query: explain
-select value, sum(key + 1) as sum from src group by value limit 20
+select value, sum(key + 1) as sum from src group by value order by value limit 20
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -222,11 +222,11 @@ STAGE PLANS:
Processor Tree:
ListSink
-PREHOOK: query: select value, sum(key + 1) as sum from src group by value limit 20
+PREHOOK: query: select value, sum(key + 1) as sum from src group by value order by value limit 20
PREHOOK: type: QUERY
PREHOOK: Input: default@src
#### A masked pattern was here ####
-POSTHOOK: query: select value, sum(key + 1) as sum from src group by value limit 20
+POSTHOOK: query: select value, sum(key + 1) as sum from src group by value order by value limit 20
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
#### A masked pattern was here ####
@@ -432,10 +432,10 @@ POSTHOOK: Input: default@alltypesorc
-16379.0
NULL
PREHOOK: query: explain
-select ctinyint, count(distinct(cdouble)) from alltypesorc group by ctinyint limit 20
+select ctinyint, count(distinct(cdouble)) from alltypesorc group by ctinyint order by ctinyint limit 20
PREHOOK: type: QUERY
POSTHOOK: query: explain
-select ctinyint, count(distinct(cdouble)) from alltypesorc group by ctinyint limit 20
+select ctinyint, count(distinct(cdouble)) from alltypesorc group by ctinyint order by ctinyint limit 20
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -492,11 +492,11 @@ STAGE PLANS:
Processor Tree:
ListSink
-PREHOOK: query: select ctinyint, count(distinct(cdouble)) from alltypesorc group by ctinyint limit 20
+PREHOOK: query: select ctinyint, count(distinct(cdouble)) from alltypesorc group by ctinyint order by ctinyint limit 20
PREHOOK: type: QUERY
PREHOOK: Input: default@alltypesorc
#### A masked pattern was here ####
-POSTHOOK: query: select ctinyint, count(distinct(cdouble)) from alltypesorc group by ctinyint limit 20
+POSTHOOK: query: select ctinyint, count(distinct(cdouble)) from alltypesorc group by ctinyint order by ctinyint limit 20
POSTHOOK: type: QUERY
POSTHOOK: Input: default@alltypesorc
#### A masked pattern was here ####
@@ -522,11 +522,11 @@ POSTHOOK: Input: default@alltypesorc
NULL 2932
PREHOOK: query: -- multi distinct
explain
-select ctinyint, count(distinct(cstring1)), count(distinct(cstring2)) from alltypesorc group by ctinyint limit 20
+select ctinyint, count(distinct(cstring1)), count(distinct(cstring2)) from alltypesorc group by ctinyint order by ctinyint limit 20
PREHOOK: type: QUERY
POSTHOOK: query: -- multi distinct
explain
-select ctinyint, count(distinct(cstring1)), count(distinct(cstring2)) from alltypesorc group by ctinyint limit 20
+select ctinyint, count(distinct(cstring1)), count(distinct(cstring2)) from alltypesorc group by ctinyint order by ctinyint limit 20
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -583,11 +583,11 @@ STAGE PLANS:
Processor Tree:
ListSink
-PREHOOK: query: select ctinyint, count(distinct(cstring1)), count(distinct(cstring2)) from alltypesorc group by ctinyint limit 20
+PREHOOK: query: select ctinyint, count(distinct(cstring1)), count(distinct(cstring2)) from alltypesorc group by ctinyint order by ctinyint limit 20
PREHOOK: type: QUERY
PREHOOK: Input: default@alltypesorc
#### A masked pattern was here ####
-POSTHOOK: query: select ctinyint, count(distinct(cstring1)), count(distinct(cstring2)) from alltypesorc group by ctinyint limit 20
+POSTHOOK: query: select ctinyint, count(distinct(cstring1)), count(distinct(cstring2)) from alltypesorc group by ctinyint order by ctinyint limit 20
POSTHOOK: type: QUERY
POSTHOOK: Input: default@alltypesorc
#### A masked pattern was here ####
@@ -942,11 +942,11 @@ STAGE PLANS:
PREHOOK: query: -- map aggregation disabled
explain
-select value, sum(key) as sum from src group by value limit 20
+select value, sum(key) as sum from src group by value order by value limit 20
PREHOOK: type: QUERY
POSTHOOK: query: -- map aggregation disabled
explain
-select value, sum(key) as sum from src group by value limit 20
+select value, sum(key) as sum from src group by value order by value limit 20
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -998,11 +998,11 @@ STAGE PLANS:
Processor Tree:
ListSink
-PREHOOK: query: select value, sum(key) as sum from src group by value limit 20
+PREHOOK: query: select value, sum(key) as sum from src group by value order by value limit 20
PREHOOK: type: QUERY
PREHOOK: Input: default@src
#### A masked pattern was here ####
-POSTHOOK: query: select value, sum(key) as sum from src group by value limit 20
+POSTHOOK: query: select value, sum(key) as sum from src group by value order by value limit 20
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
#### A masked pattern was here ####
Modified: hive/branches/spark/ql/src/test/results/clientpositive/mapreduce2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/mapreduce2.q.out?rev=1642149&r1=1642148&r2=1642149&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/mapreduce2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/mapreduce2.q.out Thu Nov 27 13:32:48 2014
@@ -95,11 +95,11 @@ POSTHOOK: Lineage: dest1.key SCRIPT [(sr
POSTHOOK: Lineage: dest1.one SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.ten SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.value SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
-PREHOOK: query: SELECT * FROM (SELECT dest1.* FROM dest1 DISTRIBUTE BY key SORT BY key, ten, one, value) T
+PREHOOK: query: SELECT * FROM (SELECT dest1.* FROM dest1 DISTRIBUTE BY key SORT BY key, ten, one, value) T ORDER BY key
PREHOOK: type: QUERY
PREHOOK: Input: default@dest1
#### A masked pattern was here ####
-POSTHOOK: query: SELECT * FROM (SELECT dest1.* FROM dest1 DISTRIBUTE BY key SORT BY key, ten, one, value) T
+POSTHOOK: query: SELECT * FROM (SELECT dest1.* FROM dest1 DISTRIBUTE BY key SORT BY key, ten, one, value) T ORDER BY key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest1
#### A masked pattern was here ####
Modified: hive/branches/spark/ql/src/test/results/clientpositive/multi_insert_gby3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/multi_insert_gby3.q.out?rev=1642149&r1=1642148&r2=1642149&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/multi_insert_gby3.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/multi_insert_gby3.q.out Thu Nov 27 13:32:48 2014
@@ -1,8 +1,10 @@
-PREHOOK: query: create table e1 (key string, keyD double)
+PREHOOK: query: -- SORT_QUERY_RESULTS
+create table e1 (key string, keyD double)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
PREHOOK: Output: default@e1
-POSTHOOK: query: create table e1 (key string, keyD double)
+POSTHOOK: query: -- SORT_QUERY_RESULTS
+create table e1 (key string, keyD double)
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
POSTHOOK: Output: default@e1
Modified: hive/branches/spark/ql/src/test/results/clientpositive/multi_join_union.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/multi_join_union.q.out?rev=1642149&r1=1642148&r2=1642149&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/multi_join_union.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/multi_join_union.q.out Thu Nov 27 13:32:48 2014
@@ -1,9 +1,13 @@
-PREHOOK: query: CREATE TABLE src11 as SELECT * FROM src
+PREHOOK: query: -- SORT_QUERY_RESULTS
+
+CREATE TABLE src11 as SELECT * FROM src
PREHOOK: type: CREATETABLE_AS_SELECT
PREHOOK: Input: default@src
PREHOOK: Output: database:default
PREHOOK: Output: default@src11
-POSTHOOK: query: CREATE TABLE src11 as SELECT * FROM src
+POSTHOOK: query: -- SORT_QUERY_RESULTS
+
+CREATE TABLE src11 as SELECT * FROM src
POSTHOOK: type: CREATETABLE_AS_SELECT
POSTHOOK: Input: default@src
POSTHOOK: Output: database:default
Modified: hive/branches/spark/ql/src/test/results/clientpositive/ppd_outer_join3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/ppd_outer_join3.q.out?rev=1642149&r1=1642148&r2=1642149&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/ppd_outer_join3.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/ppd_outer_join3.q.out Thu Nov 27 13:32:48 2014
@@ -1,4 +1,6 @@
-PREHOOK: query: EXPLAIN
+PREHOOK: query: -- SORT_QUERY_RESULTS
+
+EXPLAIN
FROM
src a
FULL OUTER JOIN
@@ -7,7 +9,9 @@ PREHOOK: query: EXPLAIN
SELECT a.key, a.value, b.key, b.value
WHERE a.key > '10' AND a.key < '20' AND b.key > '15' AND b.key < '25'
PREHOOK: type: QUERY
-POSTHOOK: query: EXPLAIN
+POSTHOOK: query: -- SORT_QUERY_RESULTS
+
+EXPLAIN
FROM
src a
FULL OUTER JOIN
Modified: hive/branches/spark/ql/src/test/results/clientpositive/ptf_decimal.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/ptf_decimal.q.out?rev=1642149&r1=1642148&r2=1642149&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/ptf_decimal.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/ptf_decimal.q.out Thu Nov 27 13:32:48 2014
@@ -1,4 +1,6 @@
-PREHOOK: query: -- 1. aggregate functions with decimal type
+PREHOOK: query: -- SORT_QUERY_RESULTS
+
+-- 1. aggregate functions with decimal type
select p_mfgr, p_retailprice,
lead(p_retailprice) over (partition by p_mfgr ORDER BY p_name) as c1,
@@ -9,7 +11,9 @@ from part
PREHOOK: type: QUERY
PREHOOK: Input: default@part
#### A masked pattern was here ####
-POSTHOOK: query: -- 1. aggregate functions with decimal type
+POSTHOOK: query: -- SORT_QUERY_RESULTS
+
+-- 1. aggregate functions with decimal type
select p_mfgr, p_retailprice,
lead(p_retailprice) over (partition by p_mfgr ORDER BY p_name) as c1,
@@ -22,30 +26,30 @@ POSTHOOK: Input: default@part
#### A masked pattern was here ####
Manufacturer#1 1173.15 1173.15 NULL 1173.15 1173.15
Manufacturer#1 1173.15 1753.76 1173.15 1173.15 1173.15
-Manufacturer#1 1753.76 1602.59 1173.15 1173.15 1753.76
-Manufacturer#1 1602.59 1414.42 1753.76 1173.15 1602.59
Manufacturer#1 1414.42 1632.66 1602.59 1173.15 1414.42
+Manufacturer#1 1602.59 1414.42 1753.76 1173.15 1602.59
Manufacturer#1 1632.66 NULL 1414.42 1173.15 1632.66
+Manufacturer#1 1753.76 1602.59 1173.15 1173.15 1753.76
Manufacturer#2 1690.68 1800.7 NULL 1690.68 1690.68
-Manufacturer#2 1800.7 2031.98 1690.68 1690.68 1800.7
-Manufacturer#2 2031.98 1698.66 1800.7 1690.68 2031.98
Manufacturer#2 1698.66 1701.6 2031.98 1690.68 1698.66
Manufacturer#2 1701.6 NULL 1698.66 1690.68 1701.6
-Manufacturer#3 1671.68 1190.27 NULL 1671.68 1671.68
+Manufacturer#2 1800.7 2031.98 1690.68 1690.68 1800.7
+Manufacturer#2 2031.98 1698.66 1800.7 1690.68 2031.98
Manufacturer#3 1190.27 1410.39 1671.68 1671.68 1190.27
+Manufacturer#3 1337.29 NULL 1922.98 1671.68 1337.29
Manufacturer#3 1410.39 1922.98 1190.27 1671.68 1410.39
+Manufacturer#3 1671.68 1190.27 NULL 1671.68 1671.68
Manufacturer#3 1922.98 1337.29 1410.39 1671.68 1922.98
-Manufacturer#3 1337.29 NULL 1922.98 1671.68 1337.29
-Manufacturer#4 1620.67 1375.42 NULL 1620.67 1620.67
-Manufacturer#4 1375.42 1206.26 1620.67 1620.67 1375.42
Manufacturer#4 1206.26 1844.92 1375.42 1620.67 1206.26
-Manufacturer#4 1844.92 1290.35 1206.26 1620.67 1844.92
Manufacturer#4 1290.35 NULL 1844.92 1620.67 1290.35
-Manufacturer#5 1789.69 1611.66 NULL 1789.69 1789.69
-Manufacturer#5 1611.66 1788.73 1789.69 1789.69 1611.66
-Manufacturer#5 1788.73 1018.1 1611.66 1789.69 1788.73
+Manufacturer#4 1375.42 1206.26 1620.67 1620.67 1375.42
+Manufacturer#4 1620.67 1375.42 NULL 1620.67 1620.67
+Manufacturer#4 1844.92 1290.35 1206.26 1620.67 1844.92
Manufacturer#5 1018.1 1464.48 1788.73 1789.69 1018.1
Manufacturer#5 1464.48 NULL 1018.1 1789.69 1464.48
+Manufacturer#5 1611.66 1788.73 1789.69 1789.69 1611.66
+Manufacturer#5 1788.73 1018.1 1611.66 1789.69 1788.73
+Manufacturer#5 1789.69 1611.66 NULL 1789.69 1789.69
PREHOOK: query: -- 2. ranking functions with decimal type
select p_mfgr, p_retailprice,
@@ -114,32 +118,32 @@ from part
POSTHOOK: type: QUERY
POSTHOOK: Input: default@part
#### A masked pattern was here ####
-Manufacturer#1 1753.76 NULL
-Manufacturer#1 1632.66 1753.76
-Manufacturer#1 1602.59 1632.66
-Manufacturer#1 1414.42 1602.59
-Manufacturer#1 1173.15 1414.42
Manufacturer#1 1173.15 1173.15
-Manufacturer#2 2031.98 NULL
-Manufacturer#2 1800.7 2031.98
-Manufacturer#2 1701.6 1800.7
-Manufacturer#2 1698.66 1701.6
+Manufacturer#1 1173.15 1414.42
+Manufacturer#1 1414.42 1602.59
+Manufacturer#1 1602.59 1632.66
+Manufacturer#1 1632.66 1753.76
+Manufacturer#1 1753.76 NULL
Manufacturer#2 1690.68 1698.66
-Manufacturer#3 1922.98 NULL
-Manufacturer#3 1671.68 1922.98
-Manufacturer#3 1410.39 1671.68
-Manufacturer#3 1337.29 1410.39
+Manufacturer#2 1698.66 1701.6
+Manufacturer#2 1701.6 1800.7
+Manufacturer#2 1800.7 2031.98
+Manufacturer#2 2031.98 NULL
Manufacturer#3 1190.27 1337.29
-Manufacturer#4 1844.92 NULL
-Manufacturer#4 1620.67 1844.92
-Manufacturer#4 1375.42 1620.67
-Manufacturer#4 1290.35 1375.42
+Manufacturer#3 1337.29 1410.39
+Manufacturer#3 1410.39 1671.68
+Manufacturer#3 1671.68 1922.98
+Manufacturer#3 1922.98 NULL
Manufacturer#4 1206.26 1290.35
-Manufacturer#5 1789.69 NULL
-Manufacturer#5 1788.73 1789.69
-Manufacturer#5 1611.66 1788.73
-Manufacturer#5 1464.48 1611.66
+Manufacturer#4 1290.35 1375.42
+Manufacturer#4 1375.42 1620.67
+Manufacturer#4 1620.67 1844.92
+Manufacturer#4 1844.92 NULL
Manufacturer#5 1018.1 1464.48
+Manufacturer#5 1464.48 1611.66
+Manufacturer#5 1611.66 1788.73
+Manufacturer#5 1788.73 1789.69
+Manufacturer#5 1789.69 NULL
PREHOOK: query: -- 4. partition by decimal
select p_mfgr, p_retailprice,
@@ -156,29 +160,29 @@ from part
POSTHOOK: type: QUERY
POSTHOOK: Input: default@part
#### A masked pattern was here ####
-Manufacturer#5 1018.1 NULL
-Manufacturer#1 1173.15 NULL
Manufacturer#1 1173.15 1173.15
-Manufacturer#3 1190.27 NULL
-Manufacturer#4 1206.26 NULL
-Manufacturer#4 1290.35 NULL
-Manufacturer#3 1337.29 NULL
-Manufacturer#4 1375.42 NULL
-Manufacturer#3 1410.39 NULL
+Manufacturer#1 1173.15 NULL
Manufacturer#1 1414.42 NULL
-Manufacturer#5 1464.48 NULL
Manufacturer#1 1602.59 NULL
-Manufacturer#5 1611.66 NULL
-Manufacturer#4 1620.67 NULL
Manufacturer#1 1632.66 NULL
-Manufacturer#3 1671.68 NULL
+Manufacturer#1 1753.76 NULL
Manufacturer#2 1690.68 NULL
Manufacturer#2 1698.66 NULL
Manufacturer#2 1701.6 NULL
-Manufacturer#1 1753.76 NULL
-Manufacturer#5 1788.73 NULL
-Manufacturer#5 1789.69 NULL
Manufacturer#2 1800.7 NULL
-Manufacturer#4 1844.92 NULL
-Manufacturer#3 1922.98 NULL
Manufacturer#2 2031.98 NULL
+Manufacturer#3 1190.27 NULL
+Manufacturer#3 1337.29 NULL
+Manufacturer#3 1410.39 NULL
+Manufacturer#3 1671.68 NULL
+Manufacturer#3 1922.98 NULL
+Manufacturer#4 1206.26 NULL
+Manufacturer#4 1290.35 NULL
+Manufacturer#4 1375.42 NULL
+Manufacturer#4 1620.67 NULL
+Manufacturer#4 1844.92 NULL
+Manufacturer#5 1018.1 NULL
+Manufacturer#5 1464.48 NULL
+Manufacturer#5 1611.66 NULL
+Manufacturer#5 1788.73 NULL
+Manufacturer#5 1789.69 NULL
Modified: hive/branches/spark/ql/src/test/results/clientpositive/ptf_general_queries.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/ptf_general_queries.q.out?rev=1642149&r1=1642148&r2=1642149&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/ptf_general_queries.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/ptf_general_queries.q.out Thu Nov 27 13:32:48 2014
@@ -1,4 +1,6 @@
-PREHOOK: query: -- 1. testNoPTFNoWindowing
+PREHOOK: query: -- SORT_QUERY_RESULTS
+
+-- 1. testNoPTFNoWindowing
select p_mfgr, p_name, p_size
from part
distribute by p_mfgr
@@ -6,7 +8,9 @@ sort by p_name
PREHOOK: type: QUERY
PREHOOK: Input: default@part
#### A masked pattern was here ####
-POSTHOOK: query: -- 1. testNoPTFNoWindowing
+POSTHOOK: query: -- SORT_QUERY_RESULTS
+
+-- 1. testNoPTFNoWindowing
select p_mfgr, p_name, p_size
from part
distribute by p_mfgr
@@ -14,31 +18,31 @@ sort by p_name
POSTHOOK: type: QUERY
POSTHOOK: Input: default@part
#### A masked pattern was here ####
-Manufacturer#5 almond antique blue firebrick mint 31
Manufacturer#1 almond antique burnished rose metallic 2
Manufacturer#1 almond antique burnished rose metallic 2
-Manufacturer#3 almond antique chartreuse khaki white 17
Manufacturer#1 almond antique chartreuse lavender yellow 34
-Manufacturer#3 almond antique forest lavender goldenrod 14
-Manufacturer#4 almond antique gainsboro frosted violet 10
-Manufacturer#5 almond antique medium spring khaki 6
-Manufacturer#3 almond antique metallic orange dim 19
-Manufacturer#3 almond antique misty red olive 1
-Manufacturer#3 almond antique olive coral navajo 45
Manufacturer#1 almond antique salmon chartreuse burlywood 6
-Manufacturer#5 almond antique sky peru orange 2
+Manufacturer#1 almond aquamarine burnished black steel 28
+Manufacturer#1 almond aquamarine pink moccasin thistle 42
Manufacturer#2 almond antique violet chocolate turquoise 14
-Manufacturer#4 almond antique violet mint lemon 39
Manufacturer#2 almond antique violet turquoise frosted 40
-Manufacturer#1 almond aquamarine burnished black steel 28
-Manufacturer#5 almond aquamarine dodger light gainsboro 46
-Manufacturer#4 almond aquamarine floral ivory bisque 27
Manufacturer#2 almond aquamarine midnight light salmon 2
-Manufacturer#1 almond aquamarine pink moccasin thistle 42
Manufacturer#2 almond aquamarine rose maroon antique 25
Manufacturer#2 almond aquamarine sandy cyan gainsboro 18
+Manufacturer#3 almond antique chartreuse khaki white 17
+Manufacturer#3 almond antique forest lavender goldenrod 14
+Manufacturer#3 almond antique metallic orange dim 19
+Manufacturer#3 almond antique misty red olive 1
+Manufacturer#3 almond antique olive coral navajo 45
+Manufacturer#4 almond antique gainsboro frosted violet 10
+Manufacturer#4 almond antique violet mint lemon 39
+Manufacturer#4 almond aquamarine floral ivory bisque 27
Manufacturer#4 almond aquamarine yellow dodger mint 7
Manufacturer#4 almond azure aquamarine papaya violet 12
+Manufacturer#5 almond antique blue firebrick mint 31
+Manufacturer#5 almond antique medium spring khaki 6
+Manufacturer#5 almond antique sky peru orange 2
+Manufacturer#5 almond aquamarine dodger light gainsboro 46
Manufacturer#5 almond azure blanched chiffon midnight 23
PREHOOK: query: -- 2. testUDAFsNoWindowingNoPTFNoGBY
select p_mfgr,p_name, p_retailprice,
Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/count.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/count.q.out?rev=1642149&r1=1642148&r2=1642149&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/count.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/count.q.out Thu Nov 27 13:32:48 2014
@@ -1,8 +1,10 @@
-PREHOOK: query: create table abcd (a int, b int, c int, d int)
+PREHOOK: query: -- SORT_QUERY_RESULTS
+create table abcd (a int, b int, c int, d int)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
PREHOOK: Output: default@abcd
-POSTHOOK: query: create table abcd (a int, b int, c int, d int)
+POSTHOOK: query: -- SORT_QUERY_RESULTS
+create table abcd (a int, b int, c int, d int)
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
POSTHOOK: Output: default@abcd
@@ -22,13 +24,13 @@ POSTHOOK: query: select * from abcd
POSTHOOK: type: QUERY
POSTHOOK: Input: default@abcd
#### A masked pattern was here ####
-NULL 35 23 6
+10 100 45 4
+10 100 NULL 5
10 1000 50 1
100 100 10 3
-12 NULL 80 2
-10 100 NULL 5
-10 100 45 4
12 100 75 7
+12 NULL 80 2
+NULL 35 23 6
PREHOOK: query: explain select a, count(distinct b), count(distinct c), sum(d) from abcd group by a
PREHOOK: type: QUERY
POSTHOOK: query: explain select a, count(distinct b), count(distinct c), sum(d) from abcd group by a
@@ -99,10 +101,10 @@ POSTHOOK: query: select a, count(distinc
POSTHOOK: type: QUERY
POSTHOOK: Input: default@abcd
#### A masked pattern was here ####
-NULL 1 1 6
10 2 2 10
-12 1 2 9
100 1 1 3
+12 1 2 9
+NULL 1 1 6
PREHOOK: query: explain select count(1), count(*), count(a), count(b), count(c), count(d), count(distinct a), count(distinct b), count(distinct c), count(distinct d), count(distinct a,b), count(distinct b,c), count(distinct c,d), count(distinct a,d), count(distinct a,c), count(distinct b,d), count(distinct a,b,c), count(distinct b,c,d), count(distinct a,c,d), count(distinct a,b,d), count(distinct a,b,c,d) from abcd
PREHOOK: type: QUERY
POSTHOOK: query: explain select count(1), count(*), count(a), count(b), count(c), count(d), count(distinct a), count(distinct b), count(distinct c), count(distinct d), count(distinct a,b), count(distinct b,c), count(distinct c,d), count(distinct a,d), count(distinct a,c), count(distinct b,d), count(distinct a,b,c), count(distinct b,c,d), count(distinct a,c,d), count(distinct a,b,d), count(distinct a,b,c,d) from abcd
@@ -236,10 +238,10 @@ POSTHOOK: query: select a, count(distinc
POSTHOOK: type: QUERY
POSTHOOK: Input: default@abcd
#### A masked pattern was here ####
-NULL 1 1 6
10 2 2 10
-12 1 2 9
100 1 1 3
+12 1 2 9
+NULL 1 1 6
PREHOOK: query: explain select count(1), count(*), count(a), count(b), count(c), count(d), count(distinct a), count(distinct b), count(distinct c), count(distinct d), count(distinct a,b), count(distinct b,c), count(distinct c,d), count(distinct a,d), count(distinct a,c), count(distinct b,d), count(distinct a,b,c), count(distinct b,c,d), count(distinct a,c,d), count(distinct a,b,d), count(distinct a,b,c,d) from abcd
PREHOOK: type: QUERY
POSTHOOK: query: explain select count(1), count(*), count(a), count(b), count(c), count(d), count(distinct a), count(distinct b), count(distinct c), count(distinct d), count(distinct a,b), count(distinct b,c), count(distinct c,d), count(distinct a,d), count(distinct a,c), count(distinct b,d), count(distinct a,b,c), count(distinct b,c,d), count(distinct a,c,d), count(distinct a,b,d), count(distinct a,b,c,d) from abcd
Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/limit_pushdown.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/limit_pushdown.q.out?rev=1642149&r1=1642148&r2=1642149&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/limit_pushdown.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/limit_pushdown.q.out Thu Nov 27 13:32:48 2014
@@ -173,10 +173,10 @@ POSTHOOK: Input: default@src
98 val_98
98 val_98
PREHOOK: query: explain
-select value, sum(key + 1) as sum from src group by value limit 20
+select value, sum(key + 1) as sum from src group by value order by value limit 20
PREHOOK: type: QUERY
POSTHOOK: query: explain
-select value, sum(key + 1) as sum from src group by value limit 20
+select value, sum(key + 1) as sum from src group by value order by value limit 20
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -240,11 +240,11 @@ STAGE PLANS:
Processor Tree:
ListSink
-PREHOOK: query: select value, sum(key + 1) as sum from src group by value limit 20
+PREHOOK: query: select value, sum(key + 1) as sum from src group by value order by value limit 20
PREHOOK: type: QUERY
PREHOOK: Input: default@src
#### A masked pattern was here ####
-POSTHOOK: query: select value, sum(key + 1) as sum from src group by value limit 20
+POSTHOOK: query: select value, sum(key + 1) as sum from src group by value order by value limit 20
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
#### A masked pattern was here ####
@@ -462,10 +462,10 @@ POSTHOOK: Input: default@alltypesorc
-16379.0
NULL
PREHOOK: query: explain
-select ctinyint, count(distinct(cdouble)) from alltypesorc group by ctinyint limit 20
+select ctinyint, count(distinct(cdouble)) from alltypesorc group by ctinyint order by ctinyint limit 20
PREHOOK: type: QUERY
POSTHOOK: query: explain
-select ctinyint, count(distinct(cdouble)) from alltypesorc group by ctinyint limit 20
+select ctinyint, count(distinct(cdouble)) from alltypesorc group by ctinyint order by ctinyint limit 20
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -528,11 +528,11 @@ STAGE PLANS:
Processor Tree:
ListSink
-PREHOOK: query: select ctinyint, count(distinct(cdouble)) from alltypesorc group by ctinyint limit 20
+PREHOOK: query: select ctinyint, count(distinct(cdouble)) from alltypesorc group by ctinyint order by ctinyint limit 20
PREHOOK: type: QUERY
PREHOOK: Input: default@alltypesorc
#### A masked pattern was here ####
-POSTHOOK: query: select ctinyint, count(distinct(cdouble)) from alltypesorc group by ctinyint limit 20
+POSTHOOK: query: select ctinyint, count(distinct(cdouble)) from alltypesorc group by ctinyint order by ctinyint limit 20
POSTHOOK: type: QUERY
POSTHOOK: Input: default@alltypesorc
#### A masked pattern was here ####
@@ -558,11 +558,11 @@ POSTHOOK: Input: default@alltypesorc
NULL 2932
PREHOOK: query: -- multi distinct
explain
-select ctinyint, count(distinct(cstring1)), count(distinct(cstring2)) from alltypesorc group by ctinyint limit 20
+select ctinyint, count(distinct(cstring1)), count(distinct(cstring2)) from alltypesorc group by ctinyint order by ctinyint limit 20
PREHOOK: type: QUERY
POSTHOOK: query: -- multi distinct
explain
-select ctinyint, count(distinct(cstring1)), count(distinct(cstring2)) from alltypesorc group by ctinyint limit 20
+select ctinyint, count(distinct(cstring1)), count(distinct(cstring2)) from alltypesorc group by ctinyint order by ctinyint limit 20
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -625,11 +625,11 @@ STAGE PLANS:
Processor Tree:
ListSink
-PREHOOK: query: select ctinyint, count(distinct(cstring1)), count(distinct(cstring2)) from alltypesorc group by ctinyint limit 20
+PREHOOK: query: select ctinyint, count(distinct(cstring1)), count(distinct(cstring2)) from alltypesorc group by ctinyint order by ctinyint limit 20
PREHOOK: type: QUERY
PREHOOK: Input: default@alltypesorc
#### A masked pattern was here ####
-POSTHOOK: query: select ctinyint, count(distinct(cstring1)), count(distinct(cstring2)) from alltypesorc group by ctinyint limit 20
+POSTHOOK: query: select ctinyint, count(distinct(cstring1)), count(distinct(cstring2)) from alltypesorc group by ctinyint order by ctinyint limit 20
POSTHOOK: type: QUERY
POSTHOOK: Input: default@alltypesorc
#### A masked pattern was here ####
@@ -958,11 +958,11 @@ STAGE PLANS:
PREHOOK: query: -- map aggregation disabled
explain
-select value, sum(key) as sum from src group by value limit 20
+select value, sum(key) as sum from src group by value order by value limit 20
PREHOOK: type: QUERY
POSTHOOK: query: -- map aggregation disabled
explain
-select value, sum(key) as sum from src group by value limit 20
+select value, sum(key) as sum from src group by value order by value limit 20
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -1020,11 +1020,11 @@ STAGE PLANS:
Processor Tree:
ListSink
-PREHOOK: query: select value, sum(key) as sum from src group by value limit 20
+PREHOOK: query: select value, sum(key) as sum from src group by value order by value limit 20
PREHOOK: type: QUERY
PREHOOK: Input: default@src
#### A masked pattern was here ####
-POSTHOOK: query: select value, sum(key) as sum from src group by value limit 20
+POSTHOOK: query: select value, sum(key) as sum from src group by value order by value limit 20
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
#### A masked pattern was here ####
Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/mapreduce2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/mapreduce2.q.out?rev=1642149&r1=1642148&r2=1642149&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/mapreduce2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/mapreduce2.q.out Thu Nov 27 13:32:48 2014
@@ -105,11 +105,11 @@ POSTHOOK: Lineage: dest1.key SCRIPT [(sr
POSTHOOK: Lineage: dest1.one SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.ten SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.value SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
-PREHOOK: query: SELECT * FROM (SELECT dest1.* FROM dest1 DISTRIBUTE BY key SORT BY key, ten, one, value) T
+PREHOOK: query: SELECT * FROM (SELECT dest1.* FROM dest1 DISTRIBUTE BY key SORT BY key, ten, one, value) T ORDER BY key
PREHOOK: type: QUERY
PREHOOK: Input: default@dest1
#### A masked pattern was here ####
-POSTHOOK: query: SELECT * FROM (SELECT dest1.* FROM dest1 DISTRIBUTE BY key SORT BY key, ten, one, value) T
+POSTHOOK: query: SELECT * FROM (SELECT dest1.* FROM dest1 DISTRIBUTE BY key SORT BY key, ten, one, value) T ORDER BY key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest1
#### A masked pattern was here ####