You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by br...@apache.org on 2014/10/06 05:44:26 UTC
svn commit: r1629562 [16/38] - in /hive/branches/spark: ./ accumulo-handler/
beeline/ beeline/src/java/org/apache/hive/beeline/ bin/ext/ common/
common/src/java/org/apache/hadoop/hive/conf/
common/src/test/org/apache/hadoop/hive/common/type/ contrib/sr...
Modified: hive/branches/spark/ql/src/test/results/clientpositive/ppr_pushdown3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/ppr_pushdown3.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/ppr_pushdown3.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/ppr_pushdown3.q.out Mon Oct 6 03:44:13 2014
@@ -2132,21 +2132,33 @@ PREHOOK: type: QUERY
POSTHOOK: query: explain select key from srcpart
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-0 is a root stage
+ Stage-1 is a root stage
+ Stage-0 depends on stages: Stage-1
STAGE PLANS:
+ Stage: Stage-1
+ Map Reduce
+ Map Operator Tree:
+ TableScan
+ alias: srcpart
+ Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string)
+ outputColumnNames: _col0
+ Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
Stage: Stage-0
Fetch Operator
limit: -1
Processor Tree:
- TableScan
- alias: srcpart
- Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: key (type: string)
- outputColumnNames: _col0
- Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
- ListSink
+ ListSink
PREHOOK: query: select key from srcpart
PREHOOK: type: QUERY
Modified: hive/branches/spark/ql/src/test/results/clientpositive/ql_rewrite_gbtoidx.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/ql_rewrite_gbtoidx.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/ql_rewrite_gbtoidx.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/ql_rewrite_gbtoidx.q.out Mon Oct 6 03:44:13 2014
@@ -306,7 +306,6 @@ group by l_shipdate
order by l_shipdate
PREHOOK: type: QUERY
PREHOOK: Input: default@default__lineitem_lineitem_lshipdate_idx__
-PREHOOK: Input: default@lineitem
#### A masked pattern was here ####
POSTHOOK: query: select l_shipdate, count(l_shipdate)
from lineitem
@@ -314,7 +313,6 @@ group by l_shipdate
order by l_shipdate
POSTHOOK: type: QUERY
POSTHOOK: Input: default@default__lineitem_lineitem_lshipdate_idx__
-POSTHOOK: Input: default@lineitem
#### A masked pattern was here ####
1992-04-27 1
1992-07-02 1
@@ -656,7 +654,6 @@ group by year(l_shipdate), month(l_shipd
order by year, month
PREHOOK: type: QUERY
PREHOOK: Input: default@default__lineitem_lineitem_lshipdate_idx__
-PREHOOK: Input: default@lineitem
#### A masked pattern was here ####
POSTHOOK: query: select year(l_shipdate) as year,
month(l_shipdate) as month,
@@ -666,7 +663,6 @@ group by year(l_shipdate), month(l_shipd
order by year, month
POSTHOOK: type: QUERY
POSTHOOK: Input: default@default__lineitem_lineitem_lshipdate_idx__
-POSTHOOK: Input: default@lineitem
#### A masked pattern was here ####
1992 4 1
1992 7 3
@@ -2632,12 +2628,10 @@ STAGE PLANS:
PREHOOK: query: select key, count(key) from tbl group by key order by key
PREHOOK: type: QUERY
PREHOOK: Input: default@default__tbl_tbl_key_idx__
-PREHOOK: Input: default@tbl
#### A masked pattern was here ####
POSTHOOK: query: select key, count(key) from tbl group by key order by key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@default__tbl_tbl_key_idx__
-POSTHOOK: Input: default@tbl
#### A masked pattern was here ####
1 1
2 3
Modified: hive/branches/spark/ql/src/test/results/clientpositive/query_properties.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/query_properties.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/query_properties.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/query_properties.q.out Mon Oct 6 03:44:13 2014
@@ -1,6 +1,5 @@
PREHOOK: query: select * from src a join src b on a.key = b.key limit 0
PREHOOK: type: QUERY
-PREHOOK: Input: default@src
#### A masked pattern was here ####
Has Join: true
Has Group By: false
@@ -12,7 +11,6 @@ Has Distribute By: false
Has Cluster By: false
PREHOOK: query: select * from src group by src.key, src.value limit 0
PREHOOK: type: QUERY
-PREHOOK: Input: default@src
#### A masked pattern was here ####
Has Join: false
Has Group By: true
@@ -24,7 +22,6 @@ Has Distribute By: false
Has Cluster By: false
PREHOOK: query: select * from src order by src.key limit 0
PREHOOK: type: QUERY
-PREHOOK: Input: default@src
#### A masked pattern was here ####
Has Join: false
Has Group By: false
@@ -36,7 +33,6 @@ Has Distribute By: false
Has Cluster By: false
PREHOOK: query: select * from src sort by src.key limit 0
PREHOOK: type: QUERY
-PREHOOK: Input: default@src
#### A masked pattern was here ####
Has Join: false
Has Group By: false
@@ -48,7 +44,6 @@ Has Distribute By: false
Has Cluster By: false
PREHOOK: query: select a.key, sum(b.value) from src a join src b on a.key = b.key group by a.key limit 0
PREHOOK: type: QUERY
-PREHOOK: Input: default@src
#### A masked pattern was here ####
Has Join: true
Has Group By: true
@@ -60,7 +55,6 @@ Has Distribute By: false
Has Cluster By: false
PREHOOK: query: select transform(*) using 'cat' from src limit 0
PREHOOK: type: QUERY
-PREHOOK: Input: default@src
#### A masked pattern was here ####
Has Join: false
Has Group By: false
@@ -72,7 +66,6 @@ Has Distribute By: false
Has Cluster By: false
PREHOOK: query: select * from src distribute by src.key limit 0
PREHOOK: type: QUERY
-PREHOOK: Input: default@src
#### A masked pattern was here ####
Has Join: false
Has Group By: false
@@ -84,7 +77,6 @@ Has Distribute By: true
Has Cluster By: false
PREHOOK: query: select * from src cluster by src.key limit 0
PREHOOK: type: QUERY
-PREHOOK: Input: default@src
#### A masked pattern was here ####
Has Join: false
Has Group By: false
@@ -96,7 +88,6 @@ Has Distribute By: false
Has Cluster By: true
PREHOOK: query: select key, sum(value) from (select a.key as key, b.value as value from src a join src b on a.key = b.key) c group by key limit 0
PREHOOK: type: QUERY
-PREHOOK: Input: default@src
#### A masked pattern was here ####
Has Join: true
Has Group By: true
@@ -108,7 +99,6 @@ Has Distribute By: false
Has Cluster By: false
PREHOOK: query: select * from src a join src b on a.key = b.key order by a.key limit 0
PREHOOK: type: QUERY
-PREHOOK: Input: default@src
#### A masked pattern was here ####
Has Join: true
Has Group By: false
@@ -120,7 +110,6 @@ Has Distribute By: false
Has Cluster By: false
PREHOOK: query: select * from src a join src b on a.key = b.key distribute by a.key sort by a.key, b.value limit 0
PREHOOK: type: QUERY
-PREHOOK: Input: default@src
#### A masked pattern was here ####
Has Join: true
Has Group By: false
Modified: hive/branches/spark/ql/src/test/results/clientpositive/quote1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/quote1.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/quote1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/quote1.q.out Mon Oct 6 03:44:13 2014
@@ -108,24 +108,18 @@ POSTHOOK: query: EXPLAIN
SELECT `int`.`location`, `int`.`type`, `int`.`table` FROM dest1 `int` WHERE `int`.`table` = '2008-04-08'
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-0 is a root stage
+ Stage-1 is a root stage
+ Stage-0 depends on stages: Stage-1
STAGE PLANS:
+ Stage: Stage-1
+ Map Reduce
+
Stage: Stage-0
Fetch Operator
limit: -1
Processor Tree:
- TableScan
- alias: int
- Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
- Filter Operator
- predicate: (table = '2008-04-08') (type: boolean)
- Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
- Select Operator
- expressions: location (type: int), type (type: string), '2008-04-08' (type: string)
- outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
- ListSink
+ ListSink
PREHOOK: query: FROM src
INSERT OVERWRITE TABLE dest1 PARTITION(`table`='2008-04-08') SELECT src.key as `partition`, src.value as `from` WHERE src.key >= 200 and src.key < 300
Modified: hive/branches/spark/ql/src/test/results/clientpositive/quote2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/quote2.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/quote2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/quote2.q.out Mon Oct 6 03:44:13 2014
@@ -8,7 +8,6 @@ SELECT
'abc\\\\\'', "abc\\\\\"",
'abc\\\\\\', "abc\\\\\\",
'abc""""\\', "abc''''\\",
- 'mysql_%\\_\%', 'mysql\\\_\\\\\%',
"awk '{print NR\"\\t\"$0}'",
'tab\ttab', "tab\ttab"
FROM src
@@ -24,7 +23,6 @@ SELECT
'abc\\\\\'', "abc\\\\\"",
'abc\\\\\\', "abc\\\\\\",
'abc""""\\', "abc''''\\",
- 'mysql_%\\_\%', 'mysql\\\_\\\\\%',
"awk '{print NR\"\\t\"$0}'",
'tab\ttab', "tab\ttab"
FROM src
@@ -42,12 +40,12 @@ STAGE PLANS:
alias: src
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
- expressions: 'abc' (type: string), 'abc' (type: string), 'abc'' (type: string), 'abc"' (type: string), 'abc\' (type: string), 'abc\' (type: string), 'abc\'' (type: string), 'abc\"' (type: string), 'abc\\' (type: string), 'abc\\' (type: string), 'abc\\'' (type: string), 'abc\\"' (type: string), 'abc\\\' (type: string), 'abc\\\' (type: string), 'abc""""\' (type: string), 'abc''''\' (type: string), 'mysql_%\_\%' (type: string), 'mysql\\_\\\%' (type: string), 'awk '{print NR"\t"$0}'' (type: string), 'tab tab' (type: string), 'tab tab' (type: string)
- outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20
- Statistics: Num rows: 500 Data size: 952500 Basic stats: COMPLETE Column stats: COMPLETE
+ expressions: 'abc' (type: string), 'abc' (type: string), 'abc'' (type: string), 'abc"' (type: string), 'abc\' (type: string), 'abc\' (type: string), 'abc\'' (type: string), 'abc\"' (type: string), 'abc\\' (type: string), 'abc\\' (type: string), 'abc\\'' (type: string), 'abc\\"' (type: string), 'abc\\\' (type: string), 'abc\\\' (type: string), 'abc""""\' (type: string), 'abc''''\' (type: string), 'awk '{print NR"\t"$0}'' (type: string), 'tab tab' (type: string), 'tab tab' (type: string)
+ outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18
+ Statistics: Num rows: 500 Data size: 857000 Basic stats: COMPLETE Column stats: COMPLETE
Limit
Number of rows: 1
- Statistics: Num rows: 1 Data size: 1905 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 1 Data size: 1714 Basic stats: COMPLETE Column stats: COMPLETE
ListSink
PREHOOK: query: SELECT
@@ -59,7 +57,6 @@ PREHOOK: query: SELECT
'abc\\\\\'', "abc\\\\\"",
'abc\\\\\\', "abc\\\\\\",
'abc""""\\', "abc''''\\",
- 'mysql_%\\_\%', 'mysql\\\_\\\\\%',
"awk '{print NR\"\\t\"$0}'",
'tab\ttab', "tab\ttab"
FROM src
@@ -76,7 +73,6 @@ POSTHOOK: query: SELECT
'abc\\\\\'', "abc\\\\\"",
'abc\\\\\\', "abc\\\\\\",
'abc""""\\', "abc''''\\",
- 'mysql_%\\_\%', 'mysql\\\_\\\\\%',
"awk '{print NR\"\\t\"$0}'",
'tab\ttab', "tab\ttab"
FROM src
@@ -84,4 +80,4 @@ LIMIT 1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
#### A masked pattern was here ####
-abc abc abc' abc" abc\ abc\ abc\' abc\" abc\\ abc\\ abc\\' abc\\" abc\\\ abc\\\ abc""""\ abc''''\ mysql_%\_\% mysql\\_\\\% awk '{print NR"\t"$0}' tab tab tab tab
+abc abc abc' abc" abc\ abc\ abc\' abc\" abc\\ abc\\ abc\\' abc\\" abc\\\ abc\\\ abc""""\ abc''''\ awk '{print NR"\t"$0}' tab tab tab tab
Modified: hive/branches/spark/ql/src/test/results/clientpositive/quotedid_basic.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/quotedid_basic.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/quotedid_basic.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/quotedid_basic.q.out Mon Oct 6 03:44:13 2014
@@ -30,21 +30,33 @@ PREHOOK: type: QUERY
POSTHOOK: query: explain select `x+1`, `y&y`, `!@#$%^&*()_q` from t1
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-0 is a root stage
+ Stage-1 is a root stage
+ Stage-0 depends on stages: Stage-1
STAGE PLANS:
+ Stage: Stage-1
+ Map Reduce
+ Map Operator Tree:
+ TableScan
+ alias: t1
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+ Select Operator
+ expressions: x+1 (type: string), y&y (type: string), !@#$%^&*()_q (type: string)
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
Stage: Stage-0
Fetch Operator
limit: -1
Processor Tree:
- TableScan
- alias: t1
- Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
- Select Operator
- expressions: x+1 (type: string), y&y (type: string), !@#$%^&*()_q (type: string)
- outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
- ListSink
+ ListSink
PREHOOK: query: explain select `x+1`, `y&y`, `!@#$%^&*()_q` from t1 where `!@#$%^&*()_q` = '1'
PREHOOK: type: QUERY
Modified: hive/branches/spark/ql/src/test/results/clientpositive/regex_col.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/regex_col.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/regex_col.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/regex_col.q.out Mon Oct 6 03:44:13 2014
@@ -28,21 +28,33 @@ POSTHOOK: query: EXPLAIN
SELECT `..` FROM srcpart
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-0 is a root stage
+ Stage-1 is a root stage
+ Stage-0 depends on stages: Stage-1
STAGE PLANS:
+ Stage: Stage-1
+ Map Reduce
+ Map Operator Tree:
+ TableScan
+ alias: srcpart
+ Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: COMPLETE
+ Select Operator
+ expressions: ds (type: string), hr (type: string)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 2000 Data size: 736000 Basic stats: COMPLETE Column stats: COMPLETE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 2000 Data size: 736000 Basic stats: COMPLETE Column stats: COMPLETE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
Stage: Stage-0
Fetch Operator
limit: -1
Processor Tree:
- TableScan
- alias: srcpart
- Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: ds (type: string), hr (type: string)
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
- ListSink
+ ListSink
PREHOOK: query: EXPLAIN
SELECT srcpart.`..` FROM srcpart
@@ -51,21 +63,33 @@ POSTHOOK: query: EXPLAIN
SELECT srcpart.`..` FROM srcpart
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-0 is a root stage
+ Stage-1 is a root stage
+ Stage-0 depends on stages: Stage-1
STAGE PLANS:
+ Stage: Stage-1
+ Map Reduce
+ Map Operator Tree:
+ TableScan
+ alias: srcpart
+ Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: COMPLETE
+ Select Operator
+ expressions: ds (type: string), hr (type: string)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 2000 Data size: 736000 Basic stats: COMPLETE Column stats: COMPLETE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 2000 Data size: 736000 Basic stats: COMPLETE Column stats: COMPLETE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
Stage: Stage-0
Fetch Operator
limit: -1
Processor Tree:
- TableScan
- alias: srcpart
- Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: ds (type: string), hr (type: string)
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
- ListSink
+ ListSink
PREHOOK: query: EXPLAIN
SELECT `..` FROM srcpart a JOIN srcpart b
@@ -265,21 +289,33 @@ POSTHOOK: query: EXPLAIN
SELECT `.e.` FROM srcpart
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-0 is a root stage
+ Stage-1 is a root stage
+ Stage-0 depends on stages: Stage-1
STAGE PLANS:
+ Stage: Stage-1
+ Map Reduce
+ Map Operator Tree:
+ TableScan
+ alias: srcpart
+ Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string)
+ outputColumnNames: _col0
+ Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
Stage: Stage-0
Fetch Operator
limit: -1
Processor Tree:
- TableScan
- alias: srcpart
- Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: key (type: string)
- outputColumnNames: _col0
- Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
- ListSink
+ ListSink
PREHOOK: query: EXPLAIN
SELECT `d.*` FROM srcpart
@@ -288,21 +324,33 @@ POSTHOOK: query: EXPLAIN
SELECT `d.*` FROM srcpart
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-0 is a root stage
+ Stage-1 is a root stage
+ Stage-0 depends on stages: Stage-1
STAGE PLANS:
+ Stage: Stage-1
+ Map Reduce
+ Map Operator Tree:
+ TableScan
+ alias: srcpart
+ Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: COMPLETE
+ Select Operator
+ expressions: ds (type: string)
+ outputColumnNames: _col0
+ Statistics: Num rows: 2000 Data size: 368000 Basic stats: COMPLETE Column stats: COMPLETE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 2000 Data size: 368000 Basic stats: COMPLETE Column stats: COMPLETE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
Stage: Stage-0
Fetch Operator
limit: -1
Processor Tree:
- TableScan
- alias: srcpart
- Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: ds (type: string)
- outputColumnNames: _col0
- Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
- ListSink
+ ListSink
PREHOOK: query: EXPLAIN
SELECT `(ds)?+.+` FROM srcpart
@@ -311,21 +359,33 @@ POSTHOOK: query: EXPLAIN
SELECT `(ds)?+.+` FROM srcpart
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-0 is a root stage
+ Stage-1 is a root stage
+ Stage-0 depends on stages: Stage-1
STAGE PLANS:
+ Stage: Stage-1
+ Map Reduce
+ Map Operator Tree:
+ TableScan
+ alias: srcpart
+ Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string), value (type: string), hr (type: string)
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
Stage: Stage-0
Fetch Operator
limit: -1
Processor Tree:
- TableScan
- alias: srcpart
- Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: key (type: string), value (type: string), hr (type: string)
- outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE
- ListSink
+ ListSink
PREHOOK: query: EXPLAIN
SELECT `(ds|hr)?+.+` FROM srcpart ORDER BY key, value LIMIT 10
Modified: hive/branches/spark/ql/src/test/results/clientpositive/schemeAuthority.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/schemeAuthority.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/schemeAuthority.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/schemeAuthority.q.out Mon Oct 6 03:44:13 2014
@@ -61,8 +61,8 @@ POSTHOOK: Input: default@dynpart
POSTHOOK: Input: default@dynpart@value=0
POSTHOOK: Input: default@dynpart@value=1
#### A masked pattern was here ####
-10
20
+10
PREHOOK: query: select key from src where (key = 10) order by key
PREHOOK: type: QUERY
PREHOOK: Input: default@src
Modified: hive/branches/spark/ql/src/test/results/clientpositive/schemeAuthority2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/schemeAuthority2.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/schemeAuthority2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/schemeAuthority2.q.out Mon Oct 6 03:44:13 2014
@@ -48,6 +48,6 @@ POSTHOOK: Input: default@dynpart
POSTHOOK: Input: default@dynpart@value=0/value2=clusterA
POSTHOOK: Input: default@dynpart@value=0/value2=clusterB
#### A masked pattern was here ####
-clusterA 10
clusterB 20
+clusterA 10
#### A masked pattern was here ####
Modified: hive/branches/spark/ql/src/test/results/clientpositive/select_dummy_source.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/select_dummy_source.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/select_dummy_source.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/select_dummy_source.q.out Mon Oct 6 03:44:13 2014
@@ -5,22 +5,34 @@ POSTHOOK: query: explain
select 'a', 100
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-0 is a root stage
+ Stage-1 is a root stage
+ Stage-0 depends on stages: Stage-1
STAGE PLANS:
+ Stage: Stage-1
+ Map Reduce
+ Map Operator Tree:
+ TableScan
+ alias: _dummy_table
+ Row Limit Per Split: 1
+ Statistics: Num rows: 0 Data size: 1 Basic stats: PARTIAL Column stats: COMPLETE
+ Select Operator
+ expressions: 'a' (type: string), 100 (type: int)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 0 Data size: 1 Basic stats: PARTIAL Column stats: COMPLETE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 0 Data size: 1 Basic stats: PARTIAL Column stats: COMPLETE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
Stage: Stage-0
Fetch Operator
limit: -1
Processor Tree:
- TableScan
- alias: _dummy_table
- Row Limit Per Split: 1
- Statistics: Num rows: 0 Data size: 1 Basic stats: PARTIAL Column stats: COMPLETE
- Select Operator
- expressions: 'a' (type: string), 100 (type: int)
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 0 Data size: 1 Basic stats: PARTIAL Column stats: COMPLETE
- ListSink
+ ListSink
PREHOOK: query: select 'a', 100
PREHOOK: type: QUERY
@@ -40,22 +52,34 @@ explain
select 1 + 1
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-0 is a root stage
+ Stage-1 is a root stage
+ Stage-0 depends on stages: Stage-1
STAGE PLANS:
+ Stage: Stage-1
+ Map Reduce
+ Map Operator Tree:
+ TableScan
+ alias: _dummy_table
+ Row Limit Per Split: 1
+ Statistics: Num rows: 0 Data size: 1 Basic stats: PARTIAL Column stats: COMPLETE
+ Select Operator
+ expressions: 2 (type: int)
+ outputColumnNames: _col0
+ Statistics: Num rows: 0 Data size: 1 Basic stats: PARTIAL Column stats: COMPLETE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 0 Data size: 1 Basic stats: PARTIAL Column stats: COMPLETE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
Stage: Stage-0
Fetch Operator
limit: -1
Processor Tree:
- TableScan
- alias: _dummy_table
- Row Limit Per Split: 1
- Statistics: Num rows: 0 Data size: 1 Basic stats: PARTIAL Column stats: COMPLETE
- Select Operator
- expressions: 2 (type: int)
- outputColumnNames: _col0
- Statistics: Num rows: 0 Data size: 1 Basic stats: PARTIAL Column stats: COMPLETE
- ListSink
+ ListSink
PREHOOK: query: select 1 + 1
PREHOOK: type: QUERY
@@ -241,22 +265,34 @@ explain
select 2 + 3,x from (select 1 + 2 x) X
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-0 is a root stage
+ Stage-1 is a root stage
+ Stage-0 depends on stages: Stage-1
STAGE PLANS:
+ Stage: Stage-1
+ Map Reduce
+ Map Operator Tree:
+ TableScan
+ alias: _dummy_table
+ Row Limit Per Split: 1
+ Statistics: Num rows: 0 Data size: 1 Basic stats: PARTIAL Column stats: COMPLETE
+ Select Operator
+ expressions: 5 (type: int), (1 + 2) (type: int)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 0 Data size: 1 Basic stats: PARTIAL Column stats: COMPLETE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 0 Data size: 1 Basic stats: PARTIAL Column stats: COMPLETE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
Stage: Stage-0
Fetch Operator
limit: -1
Processor Tree:
- TableScan
- alias: _dummy_table
- Row Limit Per Split: 1
- Statistics: Num rows: 0 Data size: 1 Basic stats: PARTIAL Column stats: COMPLETE
- Select Operator
- expressions: 5 (type: int), 3 (type: int)
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 0 Data size: 1 Basic stats: PARTIAL Column stats: COMPLETE
- ListSink
+ ListSink
PREHOOK: query: select 2 + 3,x from (select 1 + 2 x) X
PREHOOK: type: QUERY
Modified: hive/branches/spark/ql/src/test/results/clientpositive/serde_user_properties.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/serde_user_properties.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/serde_user_properties.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/serde_user_properties.q.out Mon Oct 6 03:44:13 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/alter_merge_stats_orc.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/alter_merge_stats_orc.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/alter_merge_stats_orc.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/alter_merge_stats_orc.q.out Mon Oct 6 03:44:13 2014
@@ -68,11 +68,9 @@ value string
#### A masked pattern was here ####
PREHOOK: query: analyze table src_orc_merge_test_stat compute statistics noscan
PREHOOK: type: QUERY
-PREHOOK: Input: default@src_orc_merge_test_stat
PREHOOK: Output: default@src_orc_merge_test_stat
POSTHOOK: query: analyze table src_orc_merge_test_stat compute statistics noscan
POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src_orc_merge_test_stat
POSTHOOK: Output: default@src_orc_merge_test_stat
PREHOOK: query: desc formatted src_orc_merge_test_stat
PREHOOK: type: DESCTABLE
@@ -120,11 +118,9 @@ POSTHOOK: Input: default@src_orc_merge_t
POSTHOOK: Output: default@src_orc_merge_test_stat
PREHOOK: query: analyze table src_orc_merge_test_stat compute statistics noscan
PREHOOK: type: QUERY
-PREHOOK: Input: default@src_orc_merge_test_stat
PREHOOK: Output: default@src_orc_merge_test_stat
POSTHOOK: query: analyze table src_orc_merge_test_stat compute statistics noscan
POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src_orc_merge_test_stat
POSTHOOK: Output: default@src_orc_merge_test_stat
PREHOOK: query: desc formatted src_orc_merge_test_stat
PREHOOK: type: DESCTABLE
@@ -270,12 +266,10 @@ Storage Desc Params:
serialization.format 1
PREHOOK: query: analyze table src_orc_merge_test_part_stat partition(ds='2011') compute statistics noscan
PREHOOK: type: QUERY
-PREHOOK: Input: default@src_orc_merge_test_part_stat
PREHOOK: Output: default@src_orc_merge_test_part_stat
PREHOOK: Output: default@src_orc_merge_test_part_stat@ds=2011
POSTHOOK: query: analyze table src_orc_merge_test_part_stat partition(ds='2011') compute statistics noscan
POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src_orc_merge_test_part_stat
POSTHOOK: Output: default@src_orc_merge_test_part_stat
POSTHOOK: Output: default@src_orc_merge_test_part_stat@ds=2011
PREHOOK: query: desc formatted src_orc_merge_test_part_stat partition (ds='2011')
@@ -329,12 +323,10 @@ POSTHOOK: Input: default@src_orc_merge_t
POSTHOOK: Output: default@src_orc_merge_test_part_stat@ds=2011
PREHOOK: query: analyze table src_orc_merge_test_part_stat partition(ds='2011') compute statistics noscan
PREHOOK: type: QUERY
-PREHOOK: Input: default@src_orc_merge_test_part_stat
PREHOOK: Output: default@src_orc_merge_test_part_stat
PREHOOK: Output: default@src_orc_merge_test_part_stat@ds=2011
POSTHOOK: query: analyze table src_orc_merge_test_part_stat partition(ds='2011') compute statistics noscan
POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src_orc_merge_test_part_stat
POSTHOOK: Output: default@src_orc_merge_test_part_stat
POSTHOOK: Output: default@src_orc_merge_test_part_stat@ds=2011
PREHOOK: query: desc formatted src_orc_merge_test_part_stat partition (ds='2011')
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket2.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket2.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket2.q.out Mon Oct 6 03:44:13 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket3.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket3.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket3.q.out Mon Oct 6 03:44:13 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket4.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket4.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket4.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket4.q.out Mon Oct 6 03:44:13 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/column_access_stats.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/column_access_stats.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/column_access_stats.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/column_access_stats.q.out Mon Oct 6 03:44:13 2014
@@ -83,19 +83,36 @@ PREHOOK: query: -- More complicated sele
EXPLAIN SELECT key FROM (SELECT key, val FROM T1) subq1
PREHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-0 is a root stage
+ Stage-1 is a root stage
+ Stage-0 depends on stages: Stage-1
STAGE PLANS:
+ Stage: Stage-1
+ Spark
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: t1
+ Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
+ Select Operator
+ expressions: key (type: string)
+ outputColumnNames: _col0
+ Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
Stage: Stage-0
Fetch Operator
limit: -1
Processor Tree:
- TableScan
- alias: t1
- Select Operator
- expressions: key (type: string)
- outputColumnNames: _col0
- ListSink
+ ListSink
PREHOOK: query: SELECT key FROM (SELECT key, val FROM T1) subq1
PREHOOK: type: QUERY
@@ -113,19 +130,36 @@ Columns:key
PREHOOK: query: EXPLAIN SELECT k FROM (SELECT key as k, val as v FROM T1) subq1
PREHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-0 is a root stage
+ Stage-1 is a root stage
+ Stage-0 depends on stages: Stage-1
STAGE PLANS:
+ Stage: Stage-1
+ Spark
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: t1
+ Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
+ Select Operator
+ expressions: key (type: string)
+ outputColumnNames: _col0
+ Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
Stage: Stage-0
Fetch Operator
limit: -1
Processor Tree:
- TableScan
- alias: t1
- Select Operator
- expressions: key (type: string)
- outputColumnNames: _col0
- ListSink
+ ListSink
PREHOOK: query: SELECT k FROM (SELECT key as k, val as v FROM T1) subq1
PREHOOK: type: QUERY
@@ -384,22 +418,28 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: t2
+ Statistics: Num rows: 0 Data size: 20 Basic stats: PARTIAL Column stats: NONE
Filter Operator
predicate: key is not null (type: boolean)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Reduce Output Operator
key expressions: key (type: string)
sort order: +
Map-reduce partition columns: key (type: string)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Map 3
Map Operator Tree:
TableScan
alias: t1
+ Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
Filter Operator
predicate: key is not null (type: boolean)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Reduce Output Operator
key expressions: key (type: string)
sort order: +
Map-reduce partition columns: key (type: string)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Join Operator
@@ -409,11 +449,14 @@ STAGE PLANS:
0 {KEY.reducesinkkey0}
1
outputColumnNames: _col0
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: _col0 (type: string)
outputColumnNames: _col0
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
File Output Operator
compressed: false
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -498,22 +541,28 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: t2
+ Statistics: Num rows: 0 Data size: 20 Basic stats: PARTIAL Column stats: NONE
Filter Operator
predicate: ((val = 3) and key is not null) (type: boolean)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Reduce Output Operator
key expressions: key (type: string)
sort order: +
Map-reduce partition columns: key (type: string)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Map 3
Map Operator Tree:
TableScan
alias: t1
+ Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
Filter Operator
predicate: ((val = 3) and key is not null) (type: boolean)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Reduce Output Operator
key expressions: key (type: string)
sort order: +
Map-reduce partition columns: key (type: string)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Join Operator
@@ -523,11 +572,14 @@ STAGE PLANS:
0 {KEY.reducesinkkey0}
1 {KEY.reducesinkkey0}
outputColumnNames: _col0, _col5
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: _col0 (type: string), '3' (type: string), _col5 (type: string), '3' (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
File Output Operator
compressed: false
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -579,28 +631,36 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: t2
+ Statistics: Num rows: 0 Data size: 20 Basic stats: PARTIAL Column stats: NONE
Filter Operator
predicate: ((key = 6) and val is not null) (type: boolean)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: val (type: string)
outputColumnNames: _col0
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string)
sort order: +
Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Map 3
Map Operator Tree:
TableScan
alias: t1
+ Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
Filter Operator
predicate: ((key = 5) and val is not null) (type: boolean)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: val (type: string)
outputColumnNames: _col0
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string)
sort order: +
Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Join Operator
@@ -610,11 +670,14 @@ STAGE PLANS:
0 {KEY.reducesinkkey0}
1
outputColumnNames: _col0
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: _col0 (type: string)
outputColumnNames: _col0
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
File Output Operator
compressed: false
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -681,39 +744,50 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: t3
+ Statistics: Num rows: 0 Data size: 35 Basic stats: PARTIAL Column stats: NONE
Filter Operator
predicate: key is not null (type: boolean)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Reduce Output Operator
key expressions: key (type: string)
sort order: +
Map-reduce partition columns: key (type: string)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
value expressions: val (type: string)
Map 3
Map Operator Tree:
TableScan
alias: t2
+ Statistics: Num rows: 0 Data size: 20 Basic stats: PARTIAL Column stats: NONE
Filter Operator
predicate: key is not null (type: boolean)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: key (type: string)
outputColumnNames: _col0
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string)
sort order: +
Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Map 5
Map Operator Tree:
TableScan
alias: t1
+ Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
Filter Operator
predicate: key is not null (type: boolean)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: key (type: string)
outputColumnNames: _col0
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string)
sort order: +
Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Join Operator
@@ -723,11 +797,14 @@ STAGE PLANS:
0 {KEY.reducesinkkey0}
1 {KEY.reducesinkkey0} {VALUE._col0}
outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string)
outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
File Output Operator
compressed: false
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -741,13 +818,16 @@ STAGE PLANS:
0 {KEY.reducesinkkey0}
1
outputColumnNames: _col0
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: _col0 (type: string)
outputColumnNames: _col0
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string)
sort order: +
Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Stage: Stage-0
Fetch Operator
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/count.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/count.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/count.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/count.q.out Mon Oct 6 03:44:13 2014
@@ -48,18 +48,22 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: abcd
+ Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: a (type: int), b (type: int), c (type: int), d (type: int)
outputColumnNames: a, b, c, d
+ Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: count(DISTINCT b), count(DISTINCT c), sum(d)
keys: a (type: int), b (type: int), c (type: int)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
+ Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
sort order: +++
Map-reduce partition columns: _col0 (type: int)
+ Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE
value expressions: _col5 (type: bigint)
Reducer 2
Reduce Operator Tree:
@@ -68,11 +72,14 @@ STAGE PLANS:
keys: KEY._col0 (type: int)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 2 Data size: 39 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: int), _col1 (type: bigint), _col2 (type: bigint), _col3 (type: bigint)
outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 2 Data size: 39 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
+ Statistics: Num rows: 2 Data size: 39 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -115,17 +122,21 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: abcd
+ Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: a (type: int), b (type: int), c (type: int), d (type: int)
outputColumnNames: a, b, c, d
+ Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: count(1), count(), count(a), count(b), count(c), count(d), count(DISTINCT a), count(DISTINCT b), count(DISTINCT c), count(DISTINCT d), count(DISTINCT a, b), count(DISTINCT b, c), count(DISTINCT c, d), count(DISTINCT a, d), count(DISTINCT a, c), count(DISTINCT b, d), count(DISTINCT a, b, c), count(DISTINCT b, c, d), count(DISTINCT a, c, d), count(DISTINCT a, b, d), count(DISTINCT a, b, c, d)
keys: a (type: int), b (type: int), c (type: int), d (type: int)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24
+ Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), _col3 (type: int)
sort order: ++++
+ Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE
value expressions: _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: bigint)
Reducer 2
Reduce Operator Tree:
@@ -133,11 +144,14 @@ STAGE PLANS:
aggregations: count(VALUE._col0), count(VALUE._col1), count(VALUE._col2), count(VALUE._col3), count(VALUE._col4), count(VALUE._col5), count(DISTINCT KEY._col0:0._col0), count(DISTINCT KEY._col0:1._col0), count(DISTINCT KEY._col0:2._col0), count(DISTINCT KEY._col0:3._col0), count(DISTINCT KEY._col0:4._col0, KEY._col0:4._col1), count(DISTINCT KEY._col0:5._col0, KEY._col0:5._col1), count(DISTINCT KEY._col0:6._col0, KEY._col0:6._col1), count(DISTINCT KEY._col0:7._col0, KEY._col0:7._col1), count(DISTINCT KEY._col0:8._col0, KEY._col0:8._col1), count(DISTINCT KEY._col0:9._col0, KEY._col0:9._col1), count(DISTINCT KEY._col0:10._col0, KEY._col0:10._col1, KEY._col0:10._col2), count(DISTINCT KEY._col0:11._col0, KEY._col0:11._col1, KEY._col0:11._col2), count(DISTINCT KEY._col0:12._col0, KEY._col0:12._col1, KEY._col0:12._col2), count(DISTINCT KEY._col0:13._col0, KEY._col0:13._col1, KEY._col0:13._col2), count(DISTINCT KEY._col0:14._col0, KEY._col0:14._col1, KEY._col0:14._col2, KEY.
_col0:14._col3)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20
+ Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: bigint), _col1 (type: bigint), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: bigint), _col10 (type: bigint), _col11 (type: bigint), _col12 (type: bigint), _col13 (type: bigint), _col14 (type: bigint), _col15 (type: bigint), _col16 (type: bigint), _col17 (type: bigint), _col18 (type: bigint), _col19 (type: bigint), _col20 (type: bigint)
outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20
+ Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
+ Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -177,13 +191,16 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: abcd
+ Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: a (type: int), b (type: int), c (type: int), d (type: int)
outputColumnNames: a, b, c, d
+ Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: a (type: int), b (type: int), c (type: int)
sort order: +++
Map-reduce partition columns: a (type: int)
+ Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE
value expressions: d (type: int)
Reducer 2
Reduce Operator Tree:
@@ -192,11 +209,14 @@ STAGE PLANS:
keys: KEY._col0 (type: int)
mode: complete
outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 2 Data size: 39 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: int), _col1 (type: bigint), _col2 (type: bigint), _col3 (type: bigint)
outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 2 Data size: 39 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
+ Statistics: Num rows: 2 Data size: 39 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -239,23 +259,29 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: abcd
+ Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: a (type: int), b (type: int), c (type: int), d (type: int)
outputColumnNames: a, b, c, d
+ Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: a (type: int), b (type: int), c (type: int), d (type: int)
sort order: ++++
+ Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Group By Operator
aggregations: count(1), count(), count(KEY._col0:0._col0), count(KEY._col0:1._col0), count(KEY._col0:2._col0), count(KEY._col0:3._col0), count(DISTINCT KEY._col0:0._col0), count(DISTINCT KEY._col0:1._col0), count(DISTINCT KEY._col0:2._col0), count(DISTINCT KEY._col0:3._col0), count(DISTINCT KEY._col0:4._col0, KEY._col0:4._col1), count(DISTINCT KEY._col0:5._col0, KEY._col0:5._col1), count(DISTINCT KEY._col0:6._col0, KEY._col0:6._col1), count(DISTINCT KEY._col0:7._col0, KEY._col0:7._col1), count(DISTINCT KEY._col0:8._col0, KEY._col0:8._col1), count(DISTINCT KEY._col0:9._col0, KEY._col0:9._col1), count(DISTINCT KEY._col0:10._col0, KEY._col0:10._col1, KEY._col0:10._col2), count(DISTINCT KEY._col0:11._col0, KEY._col0:11._col1, KEY._col0:11._col2), count(DISTINCT KEY._col0:12._col0, KEY._col0:12._col1, KEY._col0:12._col2), count(DISTINCT KEY._col0:13._col0, KEY._col0:13._col1, KEY._col0:13._col2), count(DISTINCT KEY._col0:14._col0, KEY._col0:14._col1, KEY._col0:14._col2, K
EY._col0:14._col3)
mode: complete
outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20
+ Statistics: Num rows: 1 Data size: 168 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: bigint), _col1 (type: bigint), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: bigint), _col10 (type: bigint), _col11 (type: bigint), _col12 (type: bigint), _col13 (type: bigint), _col14 (type: bigint), _col15 (type: bigint), _col16 (type: bigint), _col17 (type: bigint), _col18 (type: bigint), _col19 (type: bigint), _col20 (type: bigint)
outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20
+ Statistics: Num rows: 1 Data size: 168 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
+ Statistics: Num rows: 1 Data size: 168 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/ctas.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/ctas.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/ctas.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/ctas.q.out Mon Oct 6 03:44:13 2014
@@ -41,31 +41,40 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: key (type: string), value (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string), _col1 (type: string)
sort order: ++
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Limit
Number of rows: 10
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string), _col1 (type: string)
sort order: ++
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
Reducer 3
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
Limit
Number of rows: 10
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -181,31 +190,40 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: key (type: string), value (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string), _col1 (type: string)
sort order: ++
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Limit
Number of rows: 10
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string), _col1 (type: string)
sort order: ++
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
Reducer 3
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
Limit
Number of rows: 10
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -321,31 +339,40 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: (key / 2) (type: double), concat(value, '_con') (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: double), _col1 (type: string)
sort order: ++
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: double), KEY.reducesinkkey1 (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Limit
Number of rows: 10
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: double), _col1 (type: string)
sort order: ++
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
Reducer 3
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: double), KEY.reducesinkkey1 (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
Limit
Number of rows: 10
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
@@ -526,31 +553,40 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: key (type: string), value (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string), _col1 (type: string)
sort order: ++
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Limit
Number of rows: 10
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string), _col1 (type: string)
sort order: ++
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
Reducer 3
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
Limit
Number of rows: 10
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -709,13 +745,16 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
GatherStats: false
Select Operator
expressions: key (type: string), value (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string), _col1 (type: string)
sort order: ++
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
tag: -1
auto parallelism: true
Path -> Alias:
@@ -773,11 +812,14 @@ STAGE PLANS:
Select Operator
expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Limit
Number of rows: 10
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string), _col1 (type: string)
sort order: ++
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
tag: -1
auto parallelism: false
Reducer 3
@@ -786,13 +828,16 @@ STAGE PLANS:
Select Operator
expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
Limit
Number of rows: 10
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
GlobalTableId: 1
#### A masked pattern was here ####
NumFilesPerFileSink: 1
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
#### A masked pattern was here ####
table:
input format: org.apache.hadoop.mapred.TextInputFormat
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/disable_merge_for_bucketing.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/disable_merge_for_bucketing.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/disable_merge_for_bucketing.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/disable_merge_for_bucketing.q.out Mon Oct 6 03:44:13 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/escape_clusterby1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/escape_clusterby1.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/escape_clusterby1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/escape_clusterby1.q.out Mon Oct 6 03:44:13 2014
@@ -21,20 +21,25 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: key (type: string), value (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string), _col1 (type: string)
sort order: ++
Map-reduce partition columns: _col0 (type: string), _col1 (type: string)
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -67,20 +72,25 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: key (type: string), value (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string), _col1 (type: string)
sort order: ++
Map-reduce partition columns: _col0 (type: string), _col1 (type: string)
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/escape_distributeby1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/escape_distributeby1.q.out?rev=1629562&r1=1629561&r2=1629562&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/escape_distributeby1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/escape_distributeby1.q.out Mon Oct 6 03:44:13 2014
@@ -21,20 +21,25 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: key (type: string), value (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
sort order:
Map-reduce partition columns: _col0 (type: string), _col1 (type: string)
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
value expressions: _col0 (type: string), _col1 (type: string)
Reducer 2
Reduce Operator Tree:
Select Operator
expressions: VALUE._col0 (type: string), VALUE._col1 (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -67,20 +72,25 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: key (type: string), value (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
sort order:
Map-reduce partition columns: _col0 (type: string), _col1 (type: string)
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
value expressions: _col0 (type: string), _col1 (type: string)
Reducer 2
Reduce Operator Tree:
Select Operator
expressions: VALUE._col0 (type: string), VALUE._col1 (type: string)
outputColumnNames: _col0, _col1
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat