You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by ga...@apache.org on 2018/02/12 18:42:17 UTC
[07/50] [abbrv] hive git commit: HIVE-18359: Extend grouping set
limits from int to long (Prasanth Jayachandran reviewed by Jesus Camacho
Rodriguez)
http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets_grouping.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets_grouping.q.out b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets_grouping.q.out
index b81a0d3..8dd5cf0 100644
--- a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets_grouping.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets_grouping.q.out
@@ -74,18 +74,18 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: []
- keys: _col0 (type: int), _col1 (type: int), 0 (type: int)
+ keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -122,7 +122,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -130,22 +130,22 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), grouping(_col2, 1) (type: int), grouping(_col2, 0) (type: int)
+ expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1) (type: bigint), grouping(_col2, 0) (type: bigint)
outputColumnNames: _col0, _col1, _col2, _col3, _col4
Select Vectorization:
className: VectorSelectOperator
native: true
projectedOutputColumnNums: [0, 1, 2, 3, 4]
- selectExpressions: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:int, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:int
+ selectExpressions: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:bigint, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:bigint
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
@@ -235,18 +235,18 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: []
- keys: _col0 (type: int), _col1 (type: int), 0 (type: int)
+ keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 24 Data size: 192 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -283,7 +283,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -291,22 +291,22 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), grouping(_col2, 1) (type: int), grouping(_col2, 0) (type: int)
+ expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1) (type: bigint), grouping(_col2, 0) (type: bigint)
outputColumnNames: _col0, _col1, _col2, _col3, _col4
Select Vectorization:
className: VectorSelectOperator
native: true
projectedOutputColumnNums: [0, 1, 2, 3, 4]
- selectExpressions: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:int, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:int
+ selectExpressions: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:bigint, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:bigint
Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
@@ -403,18 +403,18 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: []
- keys: _col0 (type: int), _col1 (type: int), 0 (type: int)
+ keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 24 Data size: 192 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -451,7 +451,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -459,11 +459,11 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE
@@ -471,7 +471,7 @@ STAGE PLANS:
Filter Vectorization:
className: VectorFilterOperator
native: true
- predicateExpression: FilterLongColEqualLongScalar(col 3:int, val 1)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:int)
+ predicateExpression: FilterLongColEqualLongScalar(col 3:bigint, val 1)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:bigint)
predicate: (grouping(_col2, 1) = 1) (type: boolean)
Statistics: Num rows: 6 Data size: 48 Basic stats: COMPLETE Column stats: NONE
Select Operator
@@ -572,18 +572,18 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: []
- keys: _col0 (type: int), _col1 (type: int), 0 (type: int)
+ keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 24 Data size: 192 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -620,7 +620,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -628,11 +628,11 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE
@@ -640,20 +640,20 @@ STAGE PLANS:
Filter Vectorization:
className: VectorFilterOperator
native: true
- predicateExpression: FilterExprOrExpr(children: FilterLongColEqualLongScalar(col 3:int, val 1)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:int), FilterLongColEqualLongScalar(col 3:int, val 1)(children: VectorUDFAdaptor(grouping(_col2, 0)) -> 3:int))
+ predicateExpression: FilterExprOrExpr(children: FilterLongColEqualLongScalar(col 3:bigint, val 1)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:bigint), FilterLongColEqualLongScalar(col 3:bigint, val 1)(children: VectorUDFAdaptor(grouping(_col2, 0)) -> 3:bigint))
predicate: ((grouping(_col2, 0) = 1) or (grouping(_col2, 1) = 1)) (type: boolean)
Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col0 (type: int), _col1 (type: int), (grouping(_col2, 1) + grouping(_col2, 0)) (type: int), CASE WHEN (((grouping(_col2, 1) + grouping(_col2, 0)) = 1)) THEN (_col0) ELSE (null) END (type: int)
+ expressions: _col0 (type: int), _col1 (type: int), (grouping(_col2, 1) + grouping(_col2, 0)) (type: bigint), CASE WHEN (((grouping(_col2, 1) + grouping(_col2, 0)) = 1)) THEN (_col0) ELSE (null) END (type: int)
outputColumnNames: _col0, _col1, _col2, _col3
Select Vectorization:
className: VectorSelectOperator
native: true
projectedOutputColumnNums: [0, 1, 5, 4]
- selectExpressions: LongColAddLongColumn(col 3:int, col 4:int)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:int, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:int) -> 5:int, IfExprColumnNull(col 3:boolean, col 0:int, null)(children: LongColEqualLongScalar(col 6:int, val 1)(children: LongColAddLongColumn(col 3:int, col 4:int)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:int, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:int) -> 6:int) -> 3:boolean, col 0:int) -> 4:int
+ selectExpressions: LongColAddLongColumn(col 3:bigint, col 4:bigint)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:bigint, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:bigint) -> 5:bigint, IfExprColumnNull(col 3:boolean, col 0:int, null)(children: LongColEqualLongScalar(col 6:bigint, val 1)(children: LongColAddLongColumn(col 3:bigint, col 4:bigint)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:bigint, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:bigint) -> 6:bigint) -> 3:boolean, col 0:int) -> 4:int
Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col2 (type: int), _col3 (type: int)
+ key expressions: _col2 (type: bigint), _col3 (type: int)
sort order: -+
Reduce Sink Vectorization:
className: VectorReduceSinkObjectHashOperator
@@ -675,12 +675,12 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 4
- dataColumns: KEY.reducesinkkey0:int, KEY.reducesinkkey1:int, VALUE._col0:int, VALUE._col1:int
+ dataColumns: KEY.reducesinkkey0:bigint, KEY.reducesinkkey1:int, VALUE._col0:int, VALUE._col1:int
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
Select Operator
- expressions: VALUE._col0 (type: int), VALUE._col1 (type: int), KEY.reducesinkkey0 (type: int)
+ expressions: VALUE._col0 (type: int), VALUE._col1 (type: int), KEY.reducesinkkey0 (type: bigint)
outputColumnNames: _col0, _col1, _col2
Select Vectorization:
className: VectorSelectOperator
@@ -778,18 +778,18 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: []
- keys: key (type: int), value (type: int), 0 (type: int)
+ keys: key (type: int), value (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -826,7 +826,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -834,22 +834,22 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), grouping(_col2, 1) (type: int), grouping(_col2, 0) (type: int)
+ expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1) (type: bigint), grouping(_col2, 0) (type: bigint)
outputColumnNames: _col0, _col1, _col2, _col3, _col4
Select Vectorization:
className: VectorSelectOperator
native: true
projectedOutputColumnNums: [0, 1, 2, 3, 4]
- selectExpressions: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:int, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:int
+ selectExpressions: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:bigint, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:bigint
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
@@ -939,18 +939,18 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: []
- keys: key (type: int), value (type: int), 0 (type: int)
+ keys: key (type: int), value (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 24 Data size: 192 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -987,7 +987,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -995,22 +995,22 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), grouping(_col2, 1) (type: int), grouping(_col2, 0) (type: int)
+ expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1) (type: bigint), grouping(_col2, 0) (type: bigint)
outputColumnNames: _col0, _col1, _col2, _col3, _col4
Select Vectorization:
className: VectorSelectOperator
native: true
projectedOutputColumnNums: [0, 1, 2, 3, 4]
- selectExpressions: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:int, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:int
+ selectExpressions: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:bigint, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:bigint
Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
@@ -1107,11 +1107,11 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: []
- keys: key (type: int), value (type: int), 0 (type: int)
+ keys: key (type: int), value (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 24 Data size: 192 Basic stats: COMPLETE Column stats: NONE
@@ -1119,13 +1119,13 @@ STAGE PLANS:
Filter Vectorization:
className: VectorFilterOperator
native: true
- predicateExpression: FilterLongColEqualLongScalar(col 3:int, val 1)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:int)
+ predicateExpression: FilterLongColEqualLongScalar(col 3:bigint, val 1)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:bigint)
predicate: (grouping(_col2, 1) = 1) (type: boolean)
Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -1162,7 +1162,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -1170,11 +1170,11 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1
Statistics: Num rows: 6 Data size: 48 Basic stats: COMPLETE Column stats: NONE
@@ -1269,11 +1269,11 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: []
- keys: key (type: int), value (type: int), 0 (type: int)
+ keys: key (type: int), value (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 24 Data size: 192 Basic stats: COMPLETE Column stats: NONE
@@ -1281,13 +1281,13 @@ STAGE PLANS:
Filter Vectorization:
className: VectorFilterOperator
native: true
- predicateExpression: FilterExprOrExpr(children: FilterLongColEqualLongScalar(col 3:int, val 1)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:int), FilterLongColEqualLongScalar(col 3:int, val 1)(children: VectorUDFAdaptor(grouping(_col2, 0)) -> 3:int))
+ predicateExpression: FilterExprOrExpr(children: FilterLongColEqualLongScalar(col 3:bigint, val 1)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:bigint), FilterLongColEqualLongScalar(col 3:bigint, val 1)(children: VectorUDFAdaptor(grouping(_col2, 0)) -> 3:bigint))
predicate: ((grouping(_col2, 0) = 1) or (grouping(_col2, 1) = 1)) (type: boolean)
Statistics: Num rows: 24 Data size: 192 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -1324,7 +1324,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -1332,30 +1332,30 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col0 (type: int), _col1 (type: int), (grouping(_col2, 1) + grouping(_col2, 0)) (type: int)
+ expressions: _col0 (type: int), _col1 (type: int), (grouping(_col2, 1) + grouping(_col2, 0)) (type: bigint)
outputColumnNames: _col0, _col1, _col2
Select Vectorization:
className: VectorSelectOperator
native: true
projectedOutputColumnNums: [0, 1, 5]
- selectExpressions: LongColAddLongColumn(col 3:int, col 4:int)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:int, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:int) -> 5:int
+ selectExpressions: LongColAddLongColumn(col 3:bigint, col 4:bigint)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:bigint, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:bigint) -> 5:bigint
Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col2 (type: int), CASE WHEN ((_col2 = 1)) THEN (_col0) END (type: int)
+ key expressions: _col2 (type: bigint), CASE WHEN ((_col2 = 1)) THEN (_col0) END (type: int)
sort order: -+
Reduce Sink Vectorization:
className: VectorReduceSinkObjectHashOperator
keyColumnNums: [5, 4]
- keyExpressions: IfExprColumnNull(col 3:boolean, col 0:int, null)(children: LongColEqualLongScalar(col 5:int, val 1) -> 3:boolean, col 0:int) -> 4:int
+ keyExpressions: IfExprColumnNull(col 3:boolean, col 0:int, null)(children: LongColEqualLongScalar(col 5:bigint, val 1) -> 3:boolean, col 0:int) -> 4:int
native: true
nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
valueColumnNums: [0, 1]
@@ -1373,12 +1373,12 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 4
- dataColumns: KEY.reducesinkkey0:int, KEY.reducesinkkey1:int, VALUE._col0:int, VALUE._col1:int
+ dataColumns: KEY.reducesinkkey0:bigint, KEY.reducesinkkey1:int, VALUE._col0:int, VALUE._col1:int
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
Select Operator
- expressions: VALUE._col0 (type: int), VALUE._col1 (type: int), KEY.reducesinkkey0 (type: int)
+ expressions: VALUE._col0 (type: int), VALUE._col1 (type: int), KEY.reducesinkkey0 (type: bigint)
outputColumnNames: _col0, _col1, _col2
Select Vectorization:
className: VectorSelectOperator
@@ -1476,18 +1476,18 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: []
- keys: key (type: int), value (type: int), 0 (type: int)
+ keys: key (type: int), value (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 24 Data size: 192 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -1524,7 +1524,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -1532,22 +1532,22 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), grouping(_col2, 1, 0) (type: int)
+ expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1, 0) (type: bigint)
outputColumnNames: _col0, _col1, _col2, _col3
Select Vectorization:
className: VectorSelectOperator
native: true
projectedOutputColumnNums: [0, 1, 2, 3]
- selectExpressions: VectorUDFAdaptor(grouping(_col2, 1, 0)) -> 3:int
+ selectExpressions: VectorUDFAdaptor(grouping(_col2, 1, 0)) -> 3:bigint
Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
@@ -1642,18 +1642,18 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: []
- keys: key (type: int), value (type: int), 0 (type: int)
+ keys: key (type: int), value (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 24 Data size: 192 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -1690,7 +1690,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -1698,22 +1698,22 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), grouping(_col2, 0, 1) (type: int)
+ expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 0, 1) (type: bigint)
outputColumnNames: _col0, _col1, _col2, _col3
Select Vectorization:
className: VectorSelectOperator
native: true
projectedOutputColumnNums: [0, 1, 2, 3]
- selectExpressions: VectorUDFAdaptor(grouping(_col2, 0, 1)) -> 3:int
+ selectExpressions: VectorUDFAdaptor(grouping(_col2, 0, 1)) -> 3:bigint
Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
@@ -1808,18 +1808,18 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: []
- keys: key (type: int), value (type: int), 0 (type: int)
+ keys: key (type: int), value (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -1856,7 +1856,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -1864,22 +1864,22 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), grouping(_col2, 1, 0) (type: int)
+ expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1, 0) (type: bigint)
outputColumnNames: _col0, _col1, _col2, _col3
Select Vectorization:
className: VectorSelectOperator
native: true
projectedOutputColumnNums: [0, 1, 2, 3]
- selectExpressions: VectorUDFAdaptor(grouping(_col2, 1, 0)) -> 3:int
+ selectExpressions: VectorUDFAdaptor(grouping(_col2, 1, 0)) -> 3:bigint
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
@@ -1969,18 +1969,18 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: []
- keys: key (type: int), value (type: int), 0 (type: int)
+ keys: key (type: int), value (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -2017,7 +2017,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -2025,22 +2025,22 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), grouping(_col2, 0, 1) (type: int)
+ expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 0, 1) (type: bigint)
outputColumnNames: _col0, _col1, _col2, _col3
Select Vectorization:
className: VectorSelectOperator
native: true
projectedOutputColumnNums: [0, 1, 2, 3]
- selectExpressions: VectorUDFAdaptor(grouping(_col2, 0, 1)) -> 3:int
+ selectExpressions: VectorUDFAdaptor(grouping(_col2, 0, 1)) -> 3:bigint
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets_limit.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets_limit.q.out b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets_limit.q.out
index e8ca06e..dee20d1 100644
--- a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets_limit.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets_limit.q.out
@@ -74,18 +74,18 @@ STAGE PLANS:
aggregators: VectorUDAFCountStar(*) -> bigint
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:int
+ keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: [0]
- keys: a (type: string), b (type: string), 0 (type: int)
+ keys: a (type: string), b (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 24 Data size: 8832 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -124,7 +124,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 4
- dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:int, VALUE._col0:bigint
+ dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:bigint, VALUE._col0:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -134,11 +134,11 @@ STAGE PLANS:
aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:string, col 1:string, col 2:int
+ keyExpressions: col 0:string, col 1:string, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: [0]
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col3
Statistics: Num rows: 12 Data size: 4416 Basic stats: COMPLETE Column stats: NONE
@@ -275,18 +275,18 @@ STAGE PLANS:
aggregators: VectorUDAFCountStar(*) -> bigint
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:int
+ keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: [0]
- keys: a (type: string), b (type: string), 0 (type: int)
+ keys: a (type: string), b (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 24 Data size: 8832 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -325,7 +325,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 4
- dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:int, VALUE._col0:bigint
+ dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:bigint, VALUE._col0:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -335,11 +335,11 @@ STAGE PLANS:
aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:string, col 1:string, col 2:int
+ keyExpressions: col 0:string, col 1:string, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: [0]
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col3
Statistics: Num rows: 12 Data size: 4416 Basic stats: COMPLETE Column stats: NONE
@@ -476,18 +476,18 @@ STAGE PLANS:
aggregators: VectorUDAFCountStar(*) -> bigint
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:int
+ keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: [0]
- keys: a (type: string), b (type: string), 0 (type: int)
+ keys: a (type: string), b (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 12 Data size: 4416 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -526,7 +526,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 4
- dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:int, VALUE._col0:bigint
+ dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:bigint, VALUE._col0:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -536,11 +536,11 @@ STAGE PLANS:
aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:string, col 1:string, col 2:int
+ keyExpressions: col 0:string, col 1:string, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: [0]
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col3
Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE
@@ -675,18 +675,18 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:string, col 1:string, col 2:string, ConstantVectorExpression(val 0) -> 4:int
+ keyExpressions: col 0:string, col 1:string, col 2:string, ConstantVectorExpression(val 0) -> 4:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: []
- keys: a (type: string), b (type: string), c (type: string), 0 (type: int)
+ keys: a (type: string), b (type: string), c (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 18 Data size: 9936 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: bigint)
sort order: ++++
- Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: int)
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2, 3]
@@ -724,7 +724,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 4
- dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:string, KEY._col3:int
+ dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:string, KEY._col3:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -732,11 +732,11 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:string, col 1:string, col 2:string, col 3:int
+ keyExpressions: col 0:string, col 1:string, col 2:string, col 3:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: []
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 9 Data size: 4968 Basic stats: COMPLETE Column stats: NONE
http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_window.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_window.q.out b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_window.q.out
index 4de6ebb..9374bcc 100644
--- a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_window.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_window.q.out
@@ -72,25 +72,25 @@ STAGE PLANS:
aggregators: VectorUDAFMaxLong(col 1:int) -> int, VectorUDAFMaxLong(col 2:int) -> int
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, ConstantVectorExpression(val 0) -> 4:int
+ keyExpressions: col 0:int, ConstantVectorExpression(val 0) -> 4:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: [0, 1]
- keys: category (type: int), 0 (type: int)
+ keys: category (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: bigint)
sort order: ++
- Map-reduce partition columns: _col0 (type: int), _col1 (type: int)
+ Map-reduce partition columns: _col0 (type: int), _col1 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1]
native: true
nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
valueColumnNums: [2, 3]
- Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE
value expressions: _col2 (type: int), _col3 (type: int)
Execution mode: vectorized, llap
LLAP IO: all inputs
@@ -121,7 +121,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 4
- dataColumns: KEY._col0:int, KEY._col1:int, VALUE._col0:int, VALUE._col1:int
+ dataColumns: KEY._col0:int, KEY._col1:bigint, VALUE._col0:int, VALUE._col1:int
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -131,14 +131,14 @@ STAGE PLANS:
aggregators: VectorUDAFMaxLong(col 2:int) -> int, VectorUDAFMaxLong(col 3:int) -> int
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:int, col 1:int
+ keyExpressions: col 0:int, col 1:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: [0, 1]
- keys: KEY._col0 (type: int), KEY._col1 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col2, _col3
- Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE
pruneGroupingSetId: true
Filter Operator
Filter Vectorization:
@@ -146,7 +146,7 @@ STAGE PLANS:
native: true
predicateExpression: FilterLongColGreaterLongScalar(col 2:int, val 0)
predicate: (_col3 > 0) (type: boolean)
- Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 1 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
key expressions: _col0 (type: int), _col3 (type: int)
sort order: ++
@@ -158,7 +158,7 @@ STAGE PLANS:
nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
partitionColumnNums: [0]
valueColumnNums: [1]
- Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 1 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE
value expressions: _col2 (type: int)
Reducer 3
Execution mode: vectorized, llap
@@ -183,7 +183,7 @@ STAGE PLANS:
className: VectorSelectOperator
native: true
projectedOutputColumnNums: [0, 2, 1]
- Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 1 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE
PTF Operator
Function definitions:
Input definition
@@ -217,7 +217,7 @@ STAGE PLANS:
outputTypes: [int, int, int, int]
partitionExpressions: [col 0:int]
streamingColumns: [3]
- Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 1 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col0 (type: int), _col2 (type: int), _col3 (type: int), rank_window_0 (type: int)
outputColumnNames: _col0, _col1, _col2, _col3