You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by ga...@apache.org on 2018/02/12 18:42:19 UTC
[09/50] [abbrv] hive git commit: HIVE-18359: Extend grouping set
limits from int to long (Prasanth Jayachandran reviewed by Jesus Camacho
Rodriguez)
http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_id2.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_id2.q.out b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_id2.q.out
index e6075c7..74f6289 100644
--- a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_id2.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_id2.q.out
@@ -73,16 +73,16 @@ STAGE PLANS:
aggregators: VectorUDAFCountStar(*) -> bigint
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: [0]
- keys: _col0 (type: int), _col1 (type: int), 0 (type: int)
+ keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: rand() (type: double)
Reduce Sink Vectorization:
@@ -123,7 +123,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 4
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int, VALUE._col0:bigint
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint, VALUE._col0:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -133,16 +133,16 @@ STAGE PLANS:
aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint
className: VectorGroupByOperator
groupByMode: PARTIALS
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: STREAMING
projectedOutputColumnNums: [0]
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: partials
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: _col0 (type: int), _col1 (type: int)
Reduce Sink Vectorization:
@@ -166,7 +166,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 4
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int, VALUE._col0:bigint
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint, VALUE._col0:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -176,16 +176,16 @@ STAGE PLANS:
aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint
className: VectorGroupByOperator
groupByMode: FINAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: STREAMING
projectedOutputColumnNums: [0]
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: final
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), _col3 (type: bigint)
+ expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), _col3 (type: bigint)
outputColumnNames: _col0, _col1, _col2, _col3
Select Vectorization:
className: VectorSelectOperator
@@ -275,16 +275,16 @@ STAGE PLANS:
aggregators: VectorUDAFCountStar(*) -> bigint
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: [0]
- keys: _col0 (type: int), _col1 (type: int), 0 (type: int)
+ keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: rand() (type: double)
Reduce Sink Vectorization:
@@ -325,7 +325,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 4
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int, VALUE._col0:bigint
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint, VALUE._col0:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -335,16 +335,16 @@ STAGE PLANS:
aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint
className: VectorGroupByOperator
groupByMode: PARTIALS
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: STREAMING
projectedOutputColumnNums: [0]
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: partials
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: _col0 (type: int), _col1 (type: int)
Reduce Sink Vectorization:
@@ -368,7 +368,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 4
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int, VALUE._col0:bigint
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint, VALUE._col0:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -378,16 +378,16 @@ STAGE PLANS:
aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint
className: VectorGroupByOperator
groupByMode: FINAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: STREAMING
projectedOutputColumnNums: [0]
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: final
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), _col3 (type: bigint)
+ expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), _col3 (type: bigint)
outputColumnNames: _col0, _col1, _col2, _col3
Select Vectorization:
className: VectorSelectOperator
@@ -487,16 +487,16 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: []
- keys: _col0 (type: int), _col1 (type: int), 0 (type: int)
+ keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: rand() (type: double)
Reduce Sink Vectorization:
@@ -536,7 +536,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -544,16 +544,16 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: PARTIALS
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: STREAMING
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: partials
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: _col0 (type: int), _col1 (type: int)
Reduce Sink Vectorization:
@@ -576,7 +576,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -584,16 +584,16 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: FINAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: STREAMING
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: final
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col2 (type: int)
+ expressions: _col2 (type: bigint)
outputColumnNames: _col2
Select Vectorization:
className: VectorSelectOperator
@@ -606,16 +606,16 @@ STAGE PLANS:
aggregators: VectorUDAFCountStar(*) -> bigint
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 2:int
+ keyExpressions: col 2:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: [0]
- keys: _col2 (type: int)
+ keys: _col2 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int)
+ key expressions: _col0 (type: bigint)
sort order: +
Map-reduce partition columns: rand() (type: double)
Reduce Sink Vectorization:
@@ -639,7 +639,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 2
- dataColumns: KEY._col0:int, VALUE._col0:bigint
+ dataColumns: KEY._col0:bigint, VALUE._col0:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -649,18 +649,18 @@ STAGE PLANS:
aggregators: VectorUDAFCountMerge(col 1:bigint) -> bigint
className: VectorGroupByOperator
groupByMode: PARTIALS
- keyExpressions: col 0:int
+ keyExpressions: col 0:bigint
native: false
vectorProcessingMode: STREAMING
projectedOutputColumnNums: [0]
- keys: KEY._col0 (type: int)
+ keys: KEY._col0 (type: bigint)
mode: partials
outputColumnNames: _col0, _col1
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int)
+ key expressions: _col0 (type: bigint)
sort order: +
- Map-reduce partition columns: _col0 (type: int)
+ Map-reduce partition columns: _col0 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkLongOperator
keyColumnNums: [0]
@@ -681,7 +681,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 2
- dataColumns: KEY._col0:int, VALUE._col0:bigint
+ dataColumns: KEY._col0:bigint, VALUE._col0:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -691,11 +691,11 @@ STAGE PLANS:
aggregators: VectorUDAFCountMerge(col 1:bigint) -> bigint
className: VectorGroupByOperator
groupByMode: FINAL
- keyExpressions: col 0:int
+ keyExpressions: col 0:bigint
native: false
vectorProcessingMode: STREAMING
projectedOutputColumnNums: [0]
- keys: KEY._col0 (type: int)
+ keys: KEY._col0 (type: bigint)
mode: final
outputColumnNames: _col0, _col1
Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE
@@ -794,16 +794,16 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: []
- keys: _col0 (type: int), _col1 (type: int), 0 (type: int)
+ keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: rand() (type: double)
Reduce Sink Vectorization:
@@ -843,7 +843,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -851,16 +851,16 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: PARTIALS
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: STREAMING
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: partials
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: _col0 (type: int), _col1 (type: int)
Reduce Sink Vectorization:
@@ -883,7 +883,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -891,16 +891,16 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: FINAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: STREAMING
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: final
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col2 (type: int)
+ expressions: _col2 (type: bigint)
outputColumnNames: _col2
Select Vectorization:
className: VectorSelectOperator
@@ -913,16 +913,16 @@ STAGE PLANS:
aggregators: VectorUDAFCountStar(*) -> bigint
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 2:int
+ keyExpressions: col 2:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: [0]
- keys: _col2 (type: int)
+ keys: _col2 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int)
+ key expressions: _col0 (type: bigint)
sort order: +
Map-reduce partition columns: rand() (type: double)
Reduce Sink Vectorization:
@@ -946,7 +946,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 2
- dataColumns: KEY._col0:int, VALUE._col0:bigint
+ dataColumns: KEY._col0:bigint, VALUE._col0:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -956,18 +956,18 @@ STAGE PLANS:
aggregators: VectorUDAFCountMerge(col 1:bigint) -> bigint
className: VectorGroupByOperator
groupByMode: PARTIALS
- keyExpressions: col 0:int
+ keyExpressions: col 0:bigint
native: false
vectorProcessingMode: STREAMING
projectedOutputColumnNums: [0]
- keys: KEY._col0 (type: int)
+ keys: KEY._col0 (type: bigint)
mode: partials
outputColumnNames: _col0, _col1
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int)
+ key expressions: _col0 (type: bigint)
sort order: +
- Map-reduce partition columns: _col0 (type: int)
+ Map-reduce partition columns: _col0 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkLongOperator
keyColumnNums: [0]
@@ -988,7 +988,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 2
- dataColumns: KEY._col0:int, VALUE._col0:bigint
+ dataColumns: KEY._col0:bigint, VALUE._col0:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -998,11 +998,11 @@ STAGE PLANS:
aggregators: VectorUDAFCountMerge(col 1:bigint) -> bigint
className: VectorGroupByOperator
groupByMode: FINAL
- keyExpressions: col 0:int
+ keyExpressions: col 0:bigint
native: false
vectorProcessingMode: STREAMING
projectedOutputColumnNums: [0]
- keys: KEY._col0 (type: int)
+ keys: KEY._col0 (type: bigint)
mode: final
outputColumnNames: _col0, _col1
Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE
@@ -1097,16 +1097,16 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: []
- keys: _col0 (type: int), _col1 (type: int), 0 (type: int)
+ keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: rand() (type: double)
Reduce Sink Vectorization:
@@ -1146,7 +1146,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -1154,16 +1154,16 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: PARTIALS
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: STREAMING
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: partials
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: _col0 (type: int), _col1 (type: int)
Reduce Sink Vectorization:
@@ -1175,7 +1175,7 @@ STAGE PLANS:
valueColumnNums: []
Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: _col0 (type: int), _col1 (type: int)
Reduce Sink Vectorization:
@@ -1198,7 +1198,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -1206,16 +1206,16 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: FINAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: STREAMING
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: final
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col2 (type: int)
+ expressions: _col2 (type: bigint)
outputColumnNames: _col0
Select Vectorization:
className: VectorSelectOperator
@@ -1223,9 +1223,9 @@ STAGE PLANS:
projectedOutputColumnNums: [2]
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int)
+ key expressions: _col0 (type: bigint)
sort order: +
- Map-reduce partition columns: _col0 (type: int)
+ Map-reduce partition columns: _col0 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkLongOperator
keyColumnNums: [2]
@@ -1240,8 +1240,8 @@ STAGE PLANS:
condition map:
Inner Join 0 to 1
keys:
- 0 _col0 (type: int)
- 1 _col0 (type: int)
+ 0 _col0 (type: bigint)
+ 1 _col0 (type: bigint)
outputColumnNames: _col0, _col1
Statistics: Num rows: 9 Data size: 79 Basic stats: COMPLETE Column stats: NONE
File Output Operator
@@ -1263,7 +1263,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -1271,16 +1271,16 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: FINAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: STREAMING
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: final
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col2 (type: int)
+ expressions: _col2 (type: bigint)
outputColumnNames: _col0
Select Vectorization:
className: VectorSelectOperator
@@ -1288,9 +1288,9 @@ STAGE PLANS:
projectedOutputColumnNums: [2]
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int)
+ key expressions: _col0 (type: bigint)
sort order: +
- Map-reduce partition columns: _col0 (type: int)
+ Map-reduce partition columns: _col0 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkLongOperator
keyColumnNums: [2]
@@ -1425,16 +1425,16 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: []
- keys: _col0 (type: int), _col1 (type: int), 0 (type: int)
+ keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: rand() (type: double)
Reduce Sink Vectorization:
@@ -1474,7 +1474,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -1482,16 +1482,16 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: PARTIALS
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: STREAMING
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: partials
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: _col0 (type: int), _col1 (type: int)
Reduce Sink Vectorization:
@@ -1503,7 +1503,7 @@ STAGE PLANS:
valueColumnNums: []
Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
Map-reduce partition columns: _col0 (type: int), _col1 (type: int)
Reduce Sink Vectorization:
@@ -1526,7 +1526,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -1534,16 +1534,16 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: FINAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: STREAMING
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: final
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col2 (type: int)
+ expressions: _col2 (type: bigint)
outputColumnNames: _col0
Select Vectorization:
className: VectorSelectOperator
@@ -1551,9 +1551,9 @@ STAGE PLANS:
projectedOutputColumnNums: [2]
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int)
+ key expressions: _col0 (type: bigint)
sort order: +
- Map-reduce partition columns: _col0 (type: int)
+ Map-reduce partition columns: _col0 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkLongOperator
keyColumnNums: [2]
@@ -1568,8 +1568,8 @@ STAGE PLANS:
condition map:
Inner Join 0 to 1
keys:
- 0 _col0 (type: int)
- 1 _col0 (type: int)
+ 0 _col0 (type: bigint)
+ 1 _col0 (type: bigint)
outputColumnNames: _col0, _col1
Statistics: Num rows: 9 Data size: 79 Basic stats: COMPLETE Column stats: NONE
File Output Operator
@@ -1591,7 +1591,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -1599,16 +1599,16 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: FINAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: STREAMING
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: final
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col2 (type: int)
+ expressions: _col2 (type: bigint)
outputColumnNames: _col0
Select Vectorization:
className: VectorSelectOperator
@@ -1616,9 +1616,9 @@ STAGE PLANS:
projectedOutputColumnNums: [2]
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int)
+ key expressions: _col0 (type: bigint)
sort order: +
- Map-reduce partition columns: _col0 (type: int)
+ Map-reduce partition columns: _col0 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkLongOperator
keyColumnNums: [2]
@@ -1746,18 +1746,18 @@ STAGE PLANS:
aggregators: VectorUDAFCountStar(*) -> bigint
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: [0]
- keys: _col0 (type: int), _col1 (type: int), 0 (type: int)
+ keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -1795,7 +1795,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 4
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int, VALUE._col0:bigint
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint, VALUE._col0:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -1805,16 +1805,16 @@ STAGE PLANS:
aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: [0]
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), _col3 (type: bigint)
+ expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), _col3 (type: bigint)
outputColumnNames: _col0, _col1, _col2, _col3
Select Vectorization:
className: VectorSelectOperator
@@ -1912,18 +1912,18 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: []
- keys: _col0 (type: int), _col1 (type: int), 0 (type: int)
+ keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -1960,7 +1960,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -1968,16 +1968,16 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col2 (type: int)
+ expressions: _col2 (type: bigint)
outputColumnNames: _col2
Select Vectorization:
className: VectorSelectOperator
@@ -1990,18 +1990,18 @@ STAGE PLANS:
aggregators: VectorUDAFCountStar(*) -> bigint
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 2:int
+ keyExpressions: col 2:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: [0]
- keys: _col2 (type: int)
+ keys: _col2 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int)
+ key expressions: _col0 (type: bigint)
sort order: +
- Map-reduce partition columns: _col0 (type: int)
+ Map-reduce partition columns: _col0 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkLongOperator
keyColumnNums: [0]
@@ -2022,7 +2022,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 2
- dataColumns: KEY._col0:int, VALUE._col0:bigint
+ dataColumns: KEY._col0:bigint, VALUE._col0:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -2032,11 +2032,11 @@ STAGE PLANS:
aggregators: VectorUDAFCountMerge(col 1:bigint) -> bigint
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:int
+ keyExpressions: col 0:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: [0]
- keys: KEY._col0 (type: int)
+ keys: KEY._col0 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1
Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE
@@ -2130,18 +2130,18 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: []
- keys: _col0 (type: int), _col1 (type: int), 0 (type: int)
+ keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -2150,9 +2150,9 @@ STAGE PLANS:
valueColumnNums: []
Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -2189,7 +2189,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -2197,16 +2197,16 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col2 (type: int)
+ expressions: _col2 (type: bigint)
outputColumnNames: _col0
Select Vectorization:
className: VectorSelectOperator
@@ -2214,9 +2214,9 @@ STAGE PLANS:
projectedOutputColumnNums: [2]
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int)
+ key expressions: _col0 (type: bigint)
sort order: +
- Map-reduce partition columns: _col0 (type: int)
+ Map-reduce partition columns: _col0 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkLongOperator
keyColumnNums: [2]
@@ -2231,8 +2231,8 @@ STAGE PLANS:
condition map:
Inner Join 0 to 1
keys:
- 0 _col0 (type: int)
- 1 _col0 (type: int)
+ 0 _col0 (type: bigint)
+ 1 _col0 (type: bigint)
outputColumnNames: _col0, _col1
Statistics: Num rows: 9 Data size: 79 Basic stats: COMPLETE Column stats: NONE
File Output Operator
@@ -2254,7 +2254,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 3
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -2262,16 +2262,16 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: []
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col2 (type: int)
+ expressions: _col2 (type: bigint)
outputColumnNames: _col0
Select Vectorization:
className: VectorSelectOperator
@@ -2279,9 +2279,9 @@ STAGE PLANS:
projectedOutputColumnNums: [2]
Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int)
+ key expressions: _col0 (type: bigint)
sort order: +
- Map-reduce partition columns: _col0 (type: int)
+ Map-reduce partition columns: _col0 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkLongOperator
keyColumnNums: [2]
http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_id3.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_id3.q.out b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_id3.q.out
index da4b81f..da4ba84 100644
--- a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_id3.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_id3.q.out
@@ -80,11 +80,11 @@ STAGE PLANS:
aggregators: VectorUDAFCountStar(*) -> bigint
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: [0]
- keys: key (type: int), value (type: int), 0 (type: int)
+ keys: key (type: int), value (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE
@@ -92,17 +92,17 @@ STAGE PLANS:
Filter Vectorization:
className: VectorFilterOperator
native: true
- predicateExpression: FilterLongColEqualLongScalar(col 2:int, val 1)
+ predicateExpression: FilterLongColEqualLongScalar(col 2:bigint, val 1)
predicate: (_col2 = 1) (type: boolean)
Statistics: Num rows: 6 Data size: 48 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), 1 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), 1 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: int), _col1 (type: int), 1 (type: int)
+ Map-reduce partition columns: _col0 (type: int), _col1 (type: int), 1 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 4]
- keyExpressions: ConstantVectorExpression(val 1) -> 4:int
+ keyExpressions: ConstantVectorExpression(val 1) -> 4:bigint
native: true
nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
valueColumnNums: [3]
@@ -137,7 +137,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 4
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int, VALUE._col0:bigint
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint, VALUE._col0:bigint
partitionColumnCount: 0
scratchColumnTypeNames: [bigint]
Reduce Operator Tree:
@@ -147,23 +147,23 @@ STAGE PLANS:
aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 1) -> 4:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 1) -> 4:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: [0]
- keys: KEY._col0 (type: int), KEY._col1 (type: int), 1 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), 1 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col3
Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE
pruneGroupingSetId: true
Select Operator
- expressions: _col0 (type: int), _col1 (type: int), 1 (type: int), _col3 (type: bigint)
+ expressions: _col0 (type: int), _col1 (type: int), 1 (type: bigint), _col3 (type: bigint)
outputColumnNames: _col0, _col1, _col2, _col3
Select Vectorization:
className: VectorSelectOperator
native: true
projectedOutputColumnNums: [0, 1, 3, 2]
- selectExpressions: ConstantVectorExpression(val 1) -> 3:int
+ selectExpressions: ConstantVectorExpression(val 1) -> 3:bigint
Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
@@ -256,18 +256,18 @@ STAGE PLANS:
aggregators: VectorUDAFCountStar(*) -> bigint
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int
+ keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: [0]
- keys: _col0 (type: int), _col1 (type: int), 0 (type: int)
+ keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -305,7 +305,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 4
- dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int, VALUE._col0:bigint
+ dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint, VALUE._col0:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -315,11 +315,11 @@ STAGE PLANS:
aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:int, col 1:int, col 2:int
+ keyExpressions: col 0:int, col 1:int, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: [0]
- keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int)
+ keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 6 Data size: 48 Basic stats: COMPLETE Column stats: NONE
@@ -327,17 +327,17 @@ STAGE PLANS:
Filter Vectorization:
className: VectorFilterOperator
native: true
- predicateExpression: FilterLongColEqualLongScalar(col 2:int, val 1)
+ predicateExpression: FilterLongColEqualLongScalar(col 2:bigint, val 1)
predicate: (_col2 = 1) (type: boolean)
Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: _col0 (type: int), _col1 (type: int), 1 (type: int), _col3 (type: bigint)
+ expressions: _col0 (type: int), _col1 (type: int), 1 (type: bigint), _col3 (type: bigint)
outputColumnNames: _col0, _col1, _col2, _col3
Select Vectorization:
className: VectorSelectOperator
native: true
projectedOutputColumnNums: [0, 1, 4, 3]
- selectExpressions: ConstantVectorExpression(val 1) -> 4:int
+ selectExpressions: ConstantVectorExpression(val 1) -> 4:bigint
Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets1.q.out b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets1.q.out
index d2b738b..137e06a 100644
--- a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets1.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets1.q.out
@@ -88,18 +88,18 @@ STAGE PLANS:
aggregators: VectorUDAFCountStar(*) -> bigint
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:int
+ keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: [0]
- keys: a (type: string), b (type: string), 0 (type: int)
+ keys: a (type: string), b (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 24 Data size: 8832 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -137,7 +137,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 4
- dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:int, VALUE._col0:bigint
+ dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:bigint, VALUE._col0:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -147,11 +147,11 @@ STAGE PLANS:
aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:string, col 1:string, col 2:int
+ keyExpressions: col 0:string, col 1:string, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: [0]
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col3
Statistics: Num rows: 12 Data size: 4416 Basic stats: COMPLETE Column stats: NONE
@@ -250,18 +250,18 @@ STAGE PLANS:
aggregators: VectorUDAFCountStar(*) -> bigint
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:int
+ keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: [0]
- keys: a (type: string), b (type: string), 0 (type: int)
+ keys: a (type: string), b (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 24 Data size: 8832 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -299,7 +299,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 4
- dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:int, VALUE._col0:bigint
+ dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:bigint, VALUE._col0:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -309,11 +309,11 @@ STAGE PLANS:
aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:string, col 1:string, col 2:int
+ keyExpressions: col 0:string, col 1:string, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: [0]
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col3
Statistics: Num rows: 12 Data size: 4416 Basic stats: COMPLETE Column stats: NONE
@@ -412,18 +412,18 @@ STAGE PLANS:
aggregators: VectorUDAFCountStar(*) -> bigint
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:int
+ keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: [0]
- keys: a (type: string), b (type: string), 0 (type: int)
+ keys: a (type: string), b (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 24 Data size: 8832 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -461,7 +461,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 4
- dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:int, VALUE._col0:bigint
+ dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:bigint, VALUE._col0:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -471,11 +471,11 @@ STAGE PLANS:
aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:string, col 1:string, col 2:int
+ keyExpressions: col 0:string, col 1:string, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: [0]
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col3
Statistics: Num rows: 12 Data size: 4416 Basic stats: COMPLETE Column stats: NONE
@@ -574,18 +574,18 @@ STAGE PLANS:
aggregators: VectorUDAFCountStar(*) -> bigint
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:int
+ keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: [0]
- keys: a (type: string), b (type: string), 0 (type: int)
+ keys: a (type: string), b (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 12 Data size: 4416 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
sort order: +++
- Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int)
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2]
@@ -623,7 +623,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 4
- dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:int, VALUE._col0:bigint
+ dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:bigint, VALUE._col0:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -633,11 +633,11 @@ STAGE PLANS:
aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:string, col 1:string, col 2:int
+ keyExpressions: col 0:string, col 1:string, col 2:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: [0]
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col3
Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE
@@ -730,18 +730,18 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: HASH
- keyExpressions: col 0:string, col 1:string, col 2:string, ConstantVectorExpression(val 0) -> 4:int
+ keyExpressions: col 0:string, col 1:string, col 2:string, ConstantVectorExpression(val 0) -> 4:bigint
native: false
vectorProcessingMode: HASH
projectedOutputColumnNums: []
- keys: a (type: string), b (type: string), c (type: string), 0 (type: int)
+ keys: a (type: string), b (type: string), c (type: string), 0 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 18 Data size: 9936 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
- key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: int)
+ key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: bigint)
sort order: ++++
- Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: int)
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: bigint)
Reduce Sink Vectorization:
className: VectorReduceSinkMultiKeyOperator
keyColumnNums: [0, 1, 2, 3]
@@ -778,7 +778,7 @@ STAGE PLANS:
vectorized: true
rowBatchContext:
dataColumnCount: 4
- dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:string, KEY._col3:int
+ dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:string, KEY._col3:bigint
partitionColumnCount: 0
scratchColumnTypeNames: []
Reduce Operator Tree:
@@ -786,11 +786,11 @@ STAGE PLANS:
Group By Vectorization:
className: VectorGroupByOperator
groupByMode: MERGEPARTIAL
- keyExpressions: col 0:string, col 1:string, col 2:string, col 3:int
+ keyExpressions: col 0:string, col 1:string, col 2:string, col 3:bigint
native: false
vectorProcessingMode: MERGE_PARTIAL
projectedOutputColumnNums: []
- keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: int)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 9 Data size: 4968 Basic stats: COMPLETE Column stats: NONE