You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by ha...@apache.org on 2018/03/25 02:26:16 UTC
[16/19] hive git commit: HIVE-18953 : Implement CHECK constraint
(Vineet Garg via Ashutosh Chauhan)
http://git-wip-us.apache.org/repos/asf/hive/blob/5b222db3/ql/src/test/results/clientpositive/llap/default_constraint.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/default_constraint.q.out b/ql/src/test/results/clientpositive/llap/default_constraint.q.out
index 72ff7af..2b4e0c9 100644
--- a/ql/src/test/results/clientpositive/llap/default_constraint.q.out
+++ b/ql/src/test/results/clientpositive/llap/default_constraint.q.out
@@ -2399,7 +2399,7 @@ STAGE PLANS:
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
- predicate: (enforce_constraint(_col0 is not null) and enforce_constraint(_col2 is not null)) (type: boolean)
+ predicate: enforce_constraint((_col0 is not null and _col2 is not null)) (type: boolean)
Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
http://git-wip-us.apache.org/repos/asf/hive/blob/5b222db3/ql/src/test/results/clientpositive/llap/enforce_constraint_notnull.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/enforce_constraint_notnull.q.out b/ql/src/test/results/clientpositive/llap/enforce_constraint_notnull.q.out
index eeb6a7a..c125882 100644
--- a/ql/src/test/results/clientpositive/llap/enforce_constraint_notnull.q.out
+++ b/ql/src/test/results/clientpositive/llap/enforce_constraint_notnull.q.out
@@ -42,7 +42,7 @@ STAGE PLANS:
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
- predicate: (enforce_constraint(_col0 is not null) and enforce_constraint(_col2 is not null)) (type: boolean)
+ predicate: enforce_constraint((_col0 is not null and _col2 is not null)) (type: boolean)
Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
@@ -160,7 +160,7 @@ STAGE PLANS:
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
- predicate: (enforce_constraint(_col0 is not null) and enforce_constraint(_col2 is not null)) (type: boolean)
+ predicate: enforce_constraint((_col0 is not null and _col2 is not null)) (type: boolean)
Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
@@ -257,15 +257,15 @@ STAGE PLANS:
alias: src
Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
- predicate: (enforce_constraint(key is not null) and enforce_constraint(value is not null)) (type: boolean)
- Statistics: Num rows: 125 Data size: 22250 Basic stats: COMPLETE Column stats: COMPLETE
+ predicate: enforce_constraint((key is not null and value is not null)) (type: boolean)
+ Statistics: Num rows: 250 Data size: 44500 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: key (type: string), value (type: string), value (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 125 Data size: 33625 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 250 Data size: 67250 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 125 Data size: 33625 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 250 Data size: 67250 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -274,7 +274,7 @@ STAGE PLANS:
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string)
outputColumnNames: a, b, c
- Statistics: Num rows: 125 Data size: 33625 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 250 Data size: 67250 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: compute_stats(a, 'hll'), compute_stats(b, 'hll'), compute_stats(c, 'hll')
mode: hash
@@ -868,15 +868,15 @@ STAGE PLANS:
alias: src
Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
- predicate: (enforce_constraint(key is not null) and enforce_constraint(value is not null)) (type: boolean)
- Statistics: Num rows: 125 Data size: 22250 Basic stats: COMPLETE Column stats: COMPLETE
+ predicate: enforce_constraint((key is not null and value is not null)) (type: boolean)
+ Statistics: Num rows: 250 Data size: 44500 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: key (type: string), value (type: string), value (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 125 Data size: 33625 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 250 Data size: 67250 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 125 Data size: 33625 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 250 Data size: 67250 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -885,7 +885,7 @@ STAGE PLANS:
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string)
outputColumnNames: a, b, c
- Statistics: Num rows: 125 Data size: 33625 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 250 Data size: 67250 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: compute_stats(a, 'hll'), compute_stats(b, 'hll'), compute_stats(c, 'hll')
mode: hash
@@ -1477,15 +1477,15 @@ STAGE PLANS:
alias: src
Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
- predicate: (enforce_constraint(key is not null) and enforce_constraint(value is not null)) (type: boolean)
- Statistics: Num rows: 125 Data size: 22250 Basic stats: COMPLETE Column stats: COMPLETE
+ predicate: enforce_constraint((key is not null and value is not null)) (type: boolean)
+ Statistics: Num rows: 250 Data size: 44500 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: key (type: string), key (type: string), value (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 125 Data size: 33125 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 250 Data size: 66250 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 125 Data size: 33125 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 250 Data size: 66250 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -1494,7 +1494,7 @@ STAGE PLANS:
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string)
outputColumnNames: a, b, c
- Statistics: Num rows: 125 Data size: 33125 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 250 Data size: 66250 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: compute_stats(a, 'hll'), compute_stats(b, 'hll'), compute_stats(c, 'hll')
mode: hash
@@ -2495,7 +2495,7 @@ STAGE PLANS:
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
- predicate: (enforce_constraint(_col1 is not null) and enforce_constraint(_col2 is not null)) (type: boolean)
+ predicate: enforce_constraint((_col1 is not null and _col2 is not null)) (type: boolean)
Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
sort order:
@@ -2627,12 +2627,12 @@ STAGE PLANS:
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 500 Data size: 103500 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
- predicate: (enforce_constraint(_col1 is not null) and enforce_constraint(_col2 is not null)) (type: boolean)
- Statistics: Num rows: 125 Data size: 25875 Basic stats: COMPLETE Column stats: COMPLETE
+ predicate: enforce_constraint((_col1 is not null and _col2 is not null)) (type: boolean)
+ Statistics: Num rows: 250 Data size: 51750 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
sort order:
Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 125 Data size: 25875 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 250 Data size: 51750 Basic stats: COMPLETE Column stats: COMPLETE
value expressions: _col0 (type: int), _col1 (type: decimal(5,2)), _col2 (type: string)
Execution mode: llap
LLAP IO: no inputs
@@ -2642,10 +2642,10 @@ STAGE PLANS:
Select Operator
expressions: VALUE._col0 (type: int), VALUE._col1 (type: decimal(5,2)), CAST( VALUE._col2 AS varchar(128)) (type: varchar(128))
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 125 Data size: 41000 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 250 Data size: 82000 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 125 Data size: 41000 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 250 Data size: 82000 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
@@ -2655,7 +2655,7 @@ STAGE PLANS:
Select Operator
expressions: _col0 (type: int), _col1 (type: decimal(5,2)), _col2 (type: varchar(128))
outputColumnNames: i, de, vc
- Statistics: Num rows: 125 Data size: 41000 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 250 Data size: 82000 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: compute_stats(i, 'hll'), compute_stats(de, 'hll'), compute_stats(vc, 'hll')
mode: hash
@@ -2763,7 +2763,7 @@ STAGE PLANS:
Number of rows: 2
Statistics: Num rows: 2 Data size: 414 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
- predicate: (enforce_constraint(_col1 is not null) and enforce_constraint(_col2 is not null)) (type: boolean)
+ predicate: enforce_constraint((_col1 is not null and _col2 is not null)) (type: boolean)
Statistics: Num rows: 1 Data size: 207 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
sort order:
@@ -2886,7 +2886,7 @@ STAGE PLANS:
Number of rows: 2
Statistics: Num rows: 2 Data size: 414 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
- predicate: (enforce_constraint(_col1 is not null) and enforce_constraint(_col2 is not null)) (type: boolean)
+ predicate: enforce_constraint((_col1 is not null and _col2 is not null)) (type: boolean)
Statistics: Num rows: 1 Data size: 207 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
sort order:
@@ -3019,7 +3019,7 @@ STAGE PLANS:
Number of rows: 2
Statistics: Num rows: 2 Data size: 414 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
- predicate: (enforce_constraint(_col1 is not null) and enforce_constraint(_col2 is not null)) (type: boolean)
+ predicate: enforce_constraint((_col1 is not null and _col2 is not null)) (type: boolean)
Statistics: Num rows: 1 Data size: 207 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
sort order:
@@ -3124,12 +3124,12 @@ STAGE PLANS:
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 500 Data size: 103500 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
- predicate: (enforce_constraint(_col1 is not null) and enforce_constraint(_col2 is not null)) (type: boolean)
- Statistics: Num rows: 125 Data size: 25875 Basic stats: COMPLETE Column stats: COMPLETE
+ predicate: enforce_constraint((_col1 is not null and _col2 is not null)) (type: boolean)
+ Statistics: Num rows: 250 Data size: 51750 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
sort order:
Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 125 Data size: 25875 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 250 Data size: 51750 Basic stats: COMPLETE Column stats: COMPLETE
value expressions: _col0 (type: int), _col1 (type: decimal(5,2)), _col2 (type: string)
Execution mode: llap
LLAP IO: no inputs
@@ -3139,10 +3139,10 @@ STAGE PLANS:
Select Operator
expressions: VALUE._col0 (type: int), VALUE._col1 (type: decimal(5,2)), CAST( VALUE._col2 AS varchar(128)) (type: varchar(128))
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 125 Data size: 41000 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 250 Data size: 82000 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 125 Data size: 41000 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 250 Data size: 82000 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
@@ -3152,7 +3152,7 @@ STAGE PLANS:
Select Operator
expressions: _col0 (type: int), _col1 (type: decimal(5,2)), _col2 (type: varchar(128))
outputColumnNames: i, de, vc
- Statistics: Num rows: 125 Data size: 41000 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 250 Data size: 82000 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: compute_stats(i, 'hll'), compute_stats(de, 'hll'), compute_stats(vc, 'hll')
mode: hash
@@ -3328,17 +3328,17 @@ STAGE PLANS:
alias: acid_uami
Statistics: Num rows: 277 Data size: 86920 Basic stats: COMPLETE Column stats: NONE
Filter Operator
- predicate: ((de = 3.14) and enforce_constraint(i is not null) and enforce_constraint(vc is not null)) (type: boolean)
- Statistics: Num rows: 1 Data size: 313 Basic stats: COMPLETE Column stats: NONE
+ predicate: ((de = 3.14) and enforce_constraint((i is not null and vc is not null))) (type: boolean)
+ Statistics: Num rows: 2 Data size: 627 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: ROW__ID (type: struct<writeid:bigint,bucketid:int,rowid:bigint>), i (type: int), vc (type: varchar(128))
outputColumnNames: _col0, _col1, _col3
- Statistics: Num rows: 1 Data size: 313 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 2 Data size: 627 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: struct<writeid:bigint,bucketid:int,rowid:bigint>)
sort order: +
Map-reduce partition columns: UDFToInteger(_col0) (type: int)
- Statistics: Num rows: 1 Data size: 313 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 2 Data size: 627 Basic stats: COMPLETE Column stats: NONE
value expressions: _col1 (type: int), _col3 (type: varchar(128))
Execution mode: llap
LLAP IO: may be used (ACID table)
@@ -3348,10 +3348,10 @@ STAGE PLANS:
Select Operator
expressions: KEY.reducesinkkey0 (type: struct<writeid:bigint,bucketid:int,rowid:bigint>), VALUE._col0 (type: int), 3.14 (type: decimal(5,2)), VALUE._col1 (type: varchar(128))
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 1 Data size: 313 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 2 Data size: 627 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 313 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 2 Data size: 627 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
@@ -3419,19 +3419,19 @@ STAGE PLANS:
alias: src
Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
- predicate: ((key < 10) and enforce_constraint(value is not null)) (type: boolean)
- Statistics: Num rows: 83 Data size: 14774 Basic stats: COMPLETE Column stats: COMPLETE
+ predicate: (key < 10) (type: boolean)
+ Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: UDFToInteger(key) (type: int), CAST( key AS decimal(5,2)) (type: decimal(5,2)), value (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 83 Data size: 17181 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 166 Data size: 34362 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
- predicate: enforce_constraint(_col0 is not null) (type: boolean)
- Statistics: Num rows: 41 Data size: 8487 Basic stats: COMPLETE Column stats: COMPLETE
+ predicate: enforce_constraint((_col0 is not null and _col2 is not null)) (type: boolean)
+ Statistics: Num rows: 83 Data size: 17181 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
sort order:
Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 41 Data size: 8487 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 83 Data size: 17181 Basic stats: COMPLETE Column stats: COMPLETE
value expressions: _col0 (type: int), _col1 (type: decimal(5,2)), _col2 (type: string)
Filter Operator
predicate: ((key < 20) and (key > 10) and enforce_constraint(value is not null)) (type: boolean)
@@ -3469,10 +3469,10 @@ STAGE PLANS:
Select Operator
expressions: VALUE._col0 (type: int), VALUE._col1 (type: decimal(5,2)), CAST( VALUE._col2 AS varchar(128)) (type: varchar(128))
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 41 Data size: 13448 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 83 Data size: 27224 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 41 Data size: 13448 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 83 Data size: 27224 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
@@ -3482,7 +3482,7 @@ STAGE PLANS:
Select Operator
expressions: _col0 (type: int), _col1 (type: decimal(5,2)), _col2 (type: varchar(128))
outputColumnNames: i, de, vc
- Statistics: Num rows: 41 Data size: 13448 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 83 Data size: 27224 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: compute_stats(i, 'hll'), compute_stats(de, 'hll'), compute_stats(vc, 'hll')
mode: hash
@@ -3609,7 +3609,7 @@ STAGE PLANS:
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
- predicate: (enforce_constraint(_col0 is not null) and enforce_constraint(_col2 is not null)) (type: boolean)
+ predicate: enforce_constraint((_col0 is not null and _col2 is not null)) (type: boolean)
Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
@@ -3741,11 +3741,11 @@ STAGE PLANS:
Number of rows: 10
Statistics: Num rows: 10 Data size: 3600 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
- predicate: (enforce_constraint(_col0 is not null) and enforce_constraint(_col2 is not null)) (type: boolean)
- Statistics: Num rows: 2 Data size: 720 Basic stats: COMPLETE Column stats: COMPLETE
+ predicate: enforce_constraint((_col0 is not null and _col2 is not null)) (type: boolean)
+ Statistics: Num rows: 5 Data size: 1800 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 2 Data size: 720 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 5 Data size: 1800 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -3754,18 +3754,18 @@ STAGE PLANS:
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string)
outputColumnNames: a, b, c, p1
- Statistics: Num rows: 2 Data size: 720 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 5 Data size: 1800 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: compute_stats(a, 'hll'), compute_stats(b, 'hll'), compute_stats(c, 'hll')
keys: p1 (type: string), 3 (type: int)
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 1 Data size: 1411 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 2 Data size: 2822 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
key expressions: _col0 (type: string), 3 (type: int)
sort order: ++
Map-reduce partition columns: _col0 (type: string), 3 (type: int)
- Statistics: Num rows: 1 Data size: 1411 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 2 Data size: 2822 Basic stats: COMPLETE Column stats: COMPLETE
value expressions: _col2 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>), _col3 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>), _col4 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>)
Reducer 3
Execution mode: llap
@@ -3775,14 +3775,14 @@ STAGE PLANS:
keys: KEY._col0 (type: string), 3 (type: int)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 1 Data size: 1411 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 2 Data size: 2822 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col2 (type: struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>), _col3 (type: struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>), _col4 (type: struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>), _col0 (type: string), 3 (type: int)
outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 1 Data size: 1411 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 2 Data size: 2822 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 1411 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 2 Data size: 2822 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -3937,11 +3937,11 @@ STAGE PLANS:
outputColumnNames: _col0, _col1, _col2, _col3, _col4
Statistics: Num rows: 500 Data size: 183000 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
- predicate: (enforce_constraint(key is not null) and enforce_constraint(value is not null)) (type: boolean)
- Statistics: Num rows: 125 Data size: 45750 Basic stats: COMPLETE Column stats: COMPLETE
+ predicate: enforce_constraint((key is not null and value is not null)) (type: boolean)
+ Statistics: Num rows: 250 Data size: 91500 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 125 Data size: 45750 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 250 Data size: 91500 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -3950,7 +3950,7 @@ STAGE PLANS:
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string)
outputColumnNames: a, b, c
- Statistics: Num rows: 125 Data size: 45750 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 250 Data size: 91500 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: compute_stats(a, 'hll'), compute_stats(b, 'hll'), compute_stats(c, 'hll')
keys: 'yesterday' (type: string), 3 (type: int)