You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by ga...@apache.org on 2018/02/12 18:42:15 UTC
[05/50] [abbrv] hive git commit: HIVE-18359: Extend grouping set
limits from int to long (Prasanth Jayachandran reviewed by Jesus Camacho
Rodriguez)
http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/spark/groupby_cube1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/groupby_cube1.q.out b/ql/src/test/results/clientpositive/spark/groupby_cube1.q.out
index fa1480e..71ccea5 100644
--- a/ql/src/test/results/clientpositive/spark/groupby_cube1.q.out
+++ b/ql/src/test/results/clientpositive/spark/groupby_cube1.q.out
@@ -42,21 +42,21 @@ STAGE PLANS:
Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: count()
- keys: key (type: string), val (type: string), 0 (type: int)
+ keys: key (type: string), val (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE
value expressions: _col3 (type: bigint)
Reducer 2
Reduce Operator Tree:
Group By Operator
aggregations: count(VALUE._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col3
Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE
@@ -107,21 +107,21 @@ STAGE PLANS:
Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: count()
- keys: key (type: string), val (type: string), 0 (type: int)
+ keys: key (type: string), val (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE
value expressions: _col3 (type: bigint)
Reducer 2
Reduce Operator Tree:
Group By Operator
aggregations: count(VALUE._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col3
Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE
@@ -198,26 +198,26 @@ STAGE PLANS:
Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: count()
- keys: _col0 (type: string), _col1 (type: string), 0 (type: int)
+ keys: _col0 (type: string), _col1 (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE
value expressions: _col3 (type: bigint)
Reducer 2
Reduce Operator Tree:
Group By Operator
aggregations: count(VALUE._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: bigint)
+ expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint), _col3 (type: bigint)
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE
File Output Operator
@@ -288,20 +288,20 @@ STAGE PLANS:
Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: count(DISTINCT val)
- keys: key (type: string), 0 (type: int), val (type: string)
+ keys: key (type: string), 0 (type: bigint), val (type: string)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: int), _col2 (type: string)
+ key expressions: _col0 (type: string), _col1 (type: bigint), _col2 (type: string)
sort order: +++
- Map-reduce partition columns: _col0 (type: string), _col1 (type: int)
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: bigint)
Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Group By Operator
aggregations: count(DISTINCT KEY._col2:0._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col2
Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE
@@ -367,12 +367,12 @@ STAGE PLANS:
Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: count()
- keys: key (type: string), val (type: string), 0 (type: int)
+ keys: key (type: string), val (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: rand() (type: double)
Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE
@@ -381,12 +381,12 @@ STAGE PLANS:
Reduce Operator Tree:
Group By Operator
aggregations: count(VALUE._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: partials
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: _col0 (type: string), _col1 (type: string)
Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE
@@ -395,7 +395,7 @@ STAGE PLANS:
Reduce Operator Tree:
Group By Operator
aggregations: count(VALUE._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: final
outputColumnNames: _col0, _col1, _col3
Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE
@@ -473,12 +473,12 @@ STAGE PLANS:
Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: count(DISTINCT val)
- keys: key (type: string), 0 (type: int), val (type: string)
+ keys: key (type: string), 0 (type: bigint), val (type: string)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: int), _col2 (type: string)
+ key expressions: _col0 (type: string), _col1 (type: bigint), _col2 (type: string)
sort order: +++
Map-reduce partition columns: _col0 (type: string)
Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE
@@ -486,12 +486,12 @@ STAGE PLANS:
Reduce Operator Tree:
Group By Operator
aggregations: count(DISTINCT KEY._col2:0._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: bigint)
mode: partials
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: bigint)
sort order: ++
Map-reduce partition columns: _col0 (type: string)
Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE
@@ -500,7 +500,7 @@ STAGE PLANS:
Reduce Operator Tree:
Group By Operator
aggregations: count(VALUE._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: bigint)
mode: final
outputColumnNames: _col0, _col2
Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE
@@ -591,12 +591,12 @@ STAGE PLANS:
Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: count(1)
- keys: key (type: string), val (type: string), 0 (type: int)
+ keys: key (type: string), val (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: rand() (type: double)
Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE
@@ -612,12 +612,12 @@ STAGE PLANS:
Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: sum(1)
- keys: key (type: string), val (type: string), 0 (type: int)
+ keys: key (type: string), val (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: rand() (type: double)
Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE
@@ -626,12 +626,12 @@ STAGE PLANS:
Reduce Operator Tree:
Group By Operator
aggregations: count(VALUE._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: partials
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: _col0 (type: string), _col1 (type: string)
Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE
@@ -640,7 +640,7 @@ STAGE PLANS:
Reduce Operator Tree:
Group By Operator
aggregations: count(VALUE._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: final
outputColumnNames: _col0, _col1, _col3
Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE
@@ -661,12 +661,12 @@ STAGE PLANS:
Reduce Operator Tree:
Group By Operator
aggregations: sum(VALUE._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: partials
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: _col0 (type: string), _col1 (type: string)
Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE
@@ -675,7 +675,7 @@ STAGE PLANS:
Reduce Operator Tree:
Group By Operator
aggregations: sum(VALUE._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: final
outputColumnNames: _col0, _col1, _col3
Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE
http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/spark/groupby_rollup1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/groupby_rollup1.q.out b/ql/src/test/results/clientpositive/spark/groupby_rollup1.q.out
index 460e7db..cf3721d 100644
--- a/ql/src/test/results/clientpositive/spark/groupby_rollup1.q.out
+++ b/ql/src/test/results/clientpositive/spark/groupby_rollup1.q.out
@@ -42,21 +42,21 @@ STAGE PLANS:
Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: count()
- keys: key (type: string), val (type: string), 0 (type: int)
+ keys: key (type: string), val (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE
value expressions: _col3 (type: bigint)
Reducer 2
Reduce Operator Tree:
Group By Operator
aggregations: count(VALUE._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col3
Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE
@@ -127,20 +127,20 @@ STAGE PLANS:
Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: count(DISTINCT val)
- keys: key (type: string), 0 (type: int), val (type: string)
+ keys: key (type: string), 0 (type: bigint), val (type: string)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: int), _col2 (type: string)
+ key expressions: _col0 (type: string), _col1 (type: bigint), _col2 (type: string)
sort order: +++
- Map-reduce partition columns: _col0 (type: string), _col1 (type: int)
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: bigint)
Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Group By Operator
aggregations: count(DISTINCT KEY._col2:0._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col2
Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE
@@ -206,12 +206,12 @@ STAGE PLANS:
Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: count()
- keys: key (type: string), val (type: string), 0 (type: int)
+ keys: key (type: string), val (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: rand() (type: double)
Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE
@@ -220,12 +220,12 @@ STAGE PLANS:
Reduce Operator Tree:
Group By Operator
aggregations: count(VALUE._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: partials
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: _col0 (type: string), _col1 (type: string)
Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE
@@ -234,7 +234,7 @@ STAGE PLANS:
Reduce Operator Tree:
Group By Operator
aggregations: count(VALUE._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: final
outputColumnNames: _col0, _col1, _col3
Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE
@@ -306,12 +306,12 @@ STAGE PLANS:
Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: count(DISTINCT val)
- keys: key (type: string), 0 (type: int), val (type: string)
+ keys: key (type: string), 0 (type: bigint), val (type: string)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: int), _col2 (type: string)
+ key expressions: _col0 (type: string), _col1 (type: bigint), _col2 (type: string)
sort order: +++
Map-reduce partition columns: _col0 (type: string)
Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE
@@ -319,12 +319,12 @@ STAGE PLANS:
Reduce Operator Tree:
Group By Operator
aggregations: count(DISTINCT KEY._col2:0._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: bigint)
mode: partials
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: bigint)
sort order: ++
Map-reduce partition columns: _col0 (type: string)
Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE
@@ -333,7 +333,7 @@ STAGE PLANS:
Reduce Operator Tree:
Group By Operator
aggregations: count(VALUE._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: bigint)
mode: final
outputColumnNames: _col0, _col2
Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE
@@ -424,12 +424,12 @@ STAGE PLANS:
Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: count(1)
- keys: key (type: string), val (type: string), 0 (type: int)
+ keys: key (type: string), val (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: rand() (type: double)
Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE
@@ -445,12 +445,12 @@ STAGE PLANS:
Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: sum(1)
- keys: key (type: string), val (type: string), 0 (type: int)
+ keys: key (type: string), val (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: rand() (type: double)
Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE
@@ -459,12 +459,12 @@ STAGE PLANS:
Reduce Operator Tree:
Group By Operator
aggregations: count(VALUE._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: partials
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: _col0 (type: string), _col1 (type: string)
Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE
@@ -473,7 +473,7 @@ STAGE PLANS:
Reduce Operator Tree:
Group By Operator
aggregations: count(VALUE._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: final
outputColumnNames: _col0, _col1, _col3
Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE
@@ -494,12 +494,12 @@ STAGE PLANS:
Reduce Operator Tree:
Group By Operator
aggregations: sum(VALUE._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: partials
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: _col0 (type: string), _col1 (type: string)
Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE
@@ -508,7 +508,7 @@ STAGE PLANS:
Reduce Operator Tree:
Group By Operator
aggregations: sum(VALUE._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: final
outputColumnNames: _col0, _col1, _col3
Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE
http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/spark/limit_pushdown2.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/limit_pushdown2.q.out b/ql/src/test/results/clientpositive/spark/limit_pushdown2.q.out
index f6f0043..3bd49b4 100644
--- a/ql/src/test/results/clientpositive/spark/limit_pushdown2.q.out
+++ b/ql/src/test/results/clientpositive/spark/limit_pushdown2.q.out
@@ -966,21 +966,21 @@ STAGE PLANS:
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: avg(_col2)
- keys: _col0 (type: string), _col1 (type: string), 0 (type: int)
+ keys: _col0 (type: string), _col1 (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 1500 Data size: 15936 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
Statistics: Num rows: 1500 Data size: 15936 Basic stats: COMPLETE Column stats: NONE
value expressions: _col3 (type: struct<count:bigint,sum:double,input:double>)
Reducer 2
Reduce Operator Tree:
Group By Operator
aggregations: avg(VALUE._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col3
Statistics: Num rows: 750 Data size: 7968 Basic stats: COMPLETE Column stats: NONE
@@ -1051,21 +1051,21 @@ STAGE PLANS:
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: avg(_col2)
- keys: _col0 (type: string), _col1 (type: string), 0 (type: int)
+ keys: _col0 (type: string), _col1 (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 1500 Data size: 15936 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
Statistics: Num rows: 1500 Data size: 15936 Basic stats: COMPLETE Column stats: NONE
value expressions: _col3 (type: struct<count:bigint,sum:double,input:double>)
Reducer 2
Reduce Operator Tree:
Group By Operator
aggregations: avg(VALUE._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col3
Statistics: Num rows: 750 Data size: 7968 Basic stats: COMPLETE Column stats: NONE
http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/tez/multi_count_distinct.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/multi_count_distinct.q.out b/ql/src/test/results/clientpositive/tez/multi_count_distinct.q.out
index bb08f16..5f5a852 100644
--- a/ql/src/test/results/clientpositive/tez/multi_count_distinct.q.out
+++ b/ql/src/test/results/clientpositive/tez/multi_count_distinct.q.out
@@ -43,14 +43,14 @@ Stage-0
PARTITION_ONLY_SHUFFLE [RS_9]
Group By Operator [GBY_8] (rows=1 width=24)
Output:["_col0","_col1","_col2"],aggregations:["count(_col0)","count(_col1)","count(_col2)"]
- Select Operator [SEL_6] (rows=13 width=97)
+ Select Operator [SEL_6] (rows=13 width=101)
Output:["_col0","_col1","_col2"]
- Group By Operator [GBY_5] (rows=13 width=97)
+ Group By Operator [GBY_5] (rows=13 width=101)
Output:["_col0","_col1","_col2","_col3"],keys:KEY._col0, KEY._col1, KEY._col2, KEY._col3
<-Map 1 [SIMPLE_EDGE]
SHUFFLE [RS_4]
PartitionCols:_col0, _col1, _col2, _col3
- Group By Operator [GBY_3] (rows=13 width=97)
+ Group By Operator [GBY_3] (rows=13 width=101)
Output:["_col0","_col1","_col2","_col3"],keys:_col0, _col1, _col2, 0
Select Operator [SEL_1] (rows=9 width=93)
Output:["_col0","_col1","_col2"]
@@ -123,14 +123,14 @@ Stage-0
PARTITION_ONLY_SHUFFLE [RS_9]
Group By Operator [GBY_8] (rows=1 width=40)
Output:["_col0","_col1","_col2","_col3","_col4"],aggregations:["count(_col0)","count(_col1)","count(_col2)","count(_col3)","count(_col4)"]
- Select Operator [SEL_6] (rows=22 width=97)
+ Select Operator [SEL_6] (rows=22 width=101)
Output:["_col0","_col1","_col2","_col3","_col4"]
- Group By Operator [GBY_5] (rows=22 width=97)
+ Group By Operator [GBY_5] (rows=22 width=101)
Output:["_col0","_col1","_col2","_col3"],keys:KEY._col0, KEY._col1, KEY._col2, KEY._col3
<-Map 1 [SIMPLE_EDGE]
SHUFFLE [RS_4]
PartitionCols:_col0, _col1, _col2, _col3
- Group By Operator [GBY_3] (rows=22 width=97)
+ Group By Operator [GBY_3] (rows=22 width=101)
Output:["_col0","_col1","_col2","_col3"],keys:_col0, _col1, _col2, 0
Select Operator [SEL_1] (rows=9 width=93)
Output:["_col0","_col1","_col2"]
http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/vector_grouping_sets.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/vector_grouping_sets.q.out b/ql/src/test/results/clientpositive/vector_grouping_sets.q.out
index f08c270..fc8290e 100644
--- a/ql/src/test/results/clientpositive/vector_grouping_sets.q.out
+++ b/ql/src/test/results/clientpositive/vector_grouping_sets.q.out
@@ -162,18 +162,18 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 1:string, ConstantVectorExpression(val 0) -> 30:int
+ keyExpressions: col 1:string, ConstantVectorExpression(val 0) -> 30:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: []
- keys: s_store_id (type: string), 0 (type: int)
+ keys: s_store_id (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1
Statistics: Num rows: 24 Data size: 51264 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: bigint)
sort order: ++
- Map-reduce partition columns: _col0 (type: string), _col1 (type: int)
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkOperator
native: false
@@ -196,7 +196,7 @@ STAGE PLANS:
enableConditionsNotMet: hive.execution.engine mr IN [tez, spark] IS false
Reduce Operator Tree:
Group By Operator
- keys: KEY._col0 (type: string), KEY._col1 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: bigint)
mode: mergepartial
outputColumnNames: _col0
Statistics: Num rows: 12 Data size: 25632 Basic stats: COMPLETE Column stats: NONE
@@ -273,18 +273,18 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 1:string, ConstantVectorExpression(val 0) -> 30:int
+ keyExpressions: col 1:string, ConstantVectorExpression(val 0) -> 30:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: []
- keys: _col0 (type: string), 0 (type: int)
+ keys: _col0 (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1
Statistics: Num rows: 24 Data size: 51264 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: bigint)
sort order: ++
- Map-reduce partition columns: _col0 (type: string), _col1 (type: int)
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkOperator
native: false
@@ -307,12 +307,12 @@ STAGE PLANS:
enableConditionsNotMet: hive.execution.engine mr IN [tez, spark] IS false
Reduce Operator Tree:
Group By Operator
- keys: KEY._col0 (type: string), KEY._col1 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1
Statistics: Num rows: 12 Data size: 25632 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col0 (type: string), _col1 (type: int)
+ expressions: _col0 (type: string), _col1 (type: bigint)
outputColumnNames: _col0, _col1
Statistics: Num rows: 12 Data size: 25632 Basic stats: COMPLETE Column stats: NONE
File Output Operator
@@ -374,24 +374,24 @@ STAGE PLANS:
outputColumnNames: _col0
Statistics: Num rows: 12 Data size: 25632 Basic stats: COMPLETE Column stats: NONE
Group By Operator
- keys: _col0 (type: string), 0 (type: int)
+ keys: _col0 (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1
Statistics: Num rows: 24 Data size: 51264 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: bigint)
sort order: ++
- Map-reduce partition columns: _col0 (type: string), _col1 (type: int)
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: bigint)
Statistics: Num rows: 24 Data size: 51264 Basic stats: COMPLETE Column stats: NONE
Execution mode: vectorized
Reduce Operator Tree:
Group By Operator
- keys: KEY._col0 (type: string), KEY._col1 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1
Statistics: Num rows: 12 Data size: 25632 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col0 (type: string), _col1 (type: int)
+ expressions: _col0 (type: string), _col1 (type: bigint)
outputColumnNames: _col0, _col1
Statistics: Num rows: 12 Data size: 25632 Basic stats: COMPLETE Column stats: NONE
File Output Operator
http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/view_cbo.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/view_cbo.q.out b/ql/src/test/results/clientpositive/view_cbo.q.out
index e10cb7a..2fafa4f 100644
--- a/ql/src/test/results/clientpositive/view_cbo.q.out
+++ b/ql/src/test/results/clientpositive/view_cbo.q.out
@@ -26,20 +26,20 @@ STAGE PLANS:
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: avg(_col2)
- keys: _col0 (type: string), _col1 (type: string), 0 (type: int)
+ keys: _col0 (type: string), _col1 (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 1500 Data size: 15936 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
Statistics: Num rows: 1500 Data size: 15936 Basic stats: COMPLETE Column stats: NONE
value expressions: _col3 (type: struct<count:bigint,sum:double,input:double>)
Reduce Operator Tree:
Group By Operator
aggregations: avg(VALUE._col0)
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col3
Statistics: Num rows: 750 Data size: 7968 Basic stats: COMPLETE Column stats: NONE