You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by xu...@apache.org on 2014/09/24 05:16:26 UTC
svn commit: r1627210 [9/14] - in /hive/branches/spark:
itests/src/test/resources/ ql/src/test/results/clientpositive/spark/
Added: hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby.q.out?rev=1627210&view=auto
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby.q.out (added)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby.q.out Wed Sep 24 03:16:25 2014
@@ -0,0 +1,719 @@
+PREHOOK: query: -- SORT_QUERY_RESULTS
+
+--HIVE-3699 Multiple insert overwrite into multiple tables query stores same results in all tables
+create table e1 (key string, count int)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@e1
+POSTHOOK: query: -- SORT_QUERY_RESULTS
+
+--HIVE-3699 Multiple insert overwrite into multiple tables query stores same results in all tables
+create table e1 (key string, count int)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@e1
+PREHOOK: query: create table e2 (key string, count int)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@e2
+POSTHOOK: query: create table e2 (key string, count int)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@e2
+PREHOOK: query: explain FROM src
+INSERT OVERWRITE TABLE e1
+ SELECT key, COUNT(*) WHERE key>450 GROUP BY key
+INSERT OVERWRITE TABLE e2
+ SELECT key, COUNT(*) WHERE key>500 GROUP BY key
+PREHOOK: type: QUERY
+POSTHOOK: query: explain FROM src
+INSERT OVERWRITE TABLE e1
+ SELECT key, COUNT(*) WHERE key>450 GROUP BY key
+INSERT OVERWRITE TABLE e2
+ SELECT key, COUNT(*) WHERE key>500 GROUP BY key
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+ Stage-2 is a root stage
+ Stage-3 depends on stages: Stage-2
+ Stage-0 depends on stages: Stage-3
+ Stage-4 depends on stages: Stage-0
+ Stage-1 depends on stages: Stage-3
+ Stage-5 depends on stages: Stage-1
+
+STAGE PLANS:
+ Stage: Stage-2
+ Spark
+ Edges:
+ Reducer 2 <- Map 1 (GROUP SORT)
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Filter Operator
+ predicate: ((key > 450) or (key > 500)) (type: boolean)
+ Statistics: Num rows: 332 Data size: 3527 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string)
+ outputColumnNames: key
+ Statistics: Num rows: 332 Data size: 3527 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: key (type: string)
+ sort order: +
+ Map-reduce partition columns: key (type: string)
+ Statistics: Num rows: 332 Data size: 3527 Basic stats: COMPLETE Column stats: NONE
+ Reducer 2
+ Reduce Operator Tree:
+ Forward
+ Statistics: Num rows: 332 Data size: 3527 Basic stats: COMPLETE Column stats: NONE
+ Filter Operator
+ predicate: (KEY._col0 > 450) (type: boolean)
+ Statistics: Num rows: 110 Data size: 1168 Basic stats: COMPLETE Column stats: NONE
+ Group By Operator
+ aggregations: count()
+ keys: KEY._col0 (type: string)
+ mode: complete
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 55 Data size: 584 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: string), UDFToInteger(_col1) (type: int)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 55 Data size: 584 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 55 Data size: 584 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e1
+ Filter Operator
+ predicate: (KEY._col0 > 500) (type: boolean)
+ Statistics: Num rows: 110 Data size: 1168 Basic stats: COMPLETE Column stats: NONE
+ Group By Operator
+ aggregations: count()
+ keys: KEY._col0 (type: string)
+ mode: complete
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 55 Data size: 584 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: string), UDFToInteger(_col1) (type: int)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 55 Data size: 584 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 55 Data size: 584 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e2
+
+ Stage: Stage-3
+ Dependency Collection
+
+ Stage: Stage-0
+ Move Operator
+ tables:
+ replace: true
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e1
+
+ Stage: Stage-4
+ Stats-Aggr Operator
+
+ Stage: Stage-1
+ Move Operator
+ tables:
+ replace: true
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e2
+
+ Stage: Stage-5
+ Stats-Aggr Operator
+
+PREHOOK: query: FROM src
+INSERT OVERWRITE TABLE e1
+ SELECT key, COUNT(*) WHERE key>450 GROUP BY key
+INSERT OVERWRITE TABLE e2
+ SELECT key, COUNT(*) WHERE key>500 GROUP BY key
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@e1
+PREHOOK: Output: default@e2
+[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to
+[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to
+POSTHOOK: query: FROM src
+INSERT OVERWRITE TABLE e1
+ SELECT key, COUNT(*) WHERE key>450 GROUP BY key
+INSERT OVERWRITE TABLE e2
+ SELECT key, COUNT(*) WHERE key>500 GROUP BY key
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@e1
+POSTHOOK: Output: default@e2
+POSTHOOK: Lineage: e1.count EXPRESSION [(src)src.null, ]
+POSTHOOK: Lineage: e1.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: e2.count EXPRESSION [(src)src.null, ]
+POSTHOOK: Lineage: e2.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+PREHOOK: query: select * from e1
+PREHOOK: type: QUERY
+PREHOOK: Input: default@e1
+#### A masked pattern was here ####
+POSTHOOK: query: select * from e1
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@e1
+#### A masked pattern was here ####
+452 1
+453 1
+454 3
+455 1
+457 1
+458 2
+459 2
+460 1
+462 2
+463 2
+466 3
+467 1
+468 4
+469 5
+470 1
+472 1
+475 1
+477 1
+478 2
+479 1
+480 3
+481 1
+482 1
+483 1
+484 1
+485 1
+487 1
+489 4
+490 1
+491 1
+492 2
+493 1
+494 1
+495 1
+496 1
+497 1
+498 3
+PREHOOK: query: select * from e2
+PREHOOK: type: QUERY
+PREHOOK: Input: default@e2
+#### A masked pattern was here ####
+POSTHOOK: query: select * from e2
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@e2
+#### A masked pattern was here ####
+PREHOOK: query: explain FROM src
+INSERT OVERWRITE TABLE e1
+ SELECT key, COUNT(*) WHERE key>450 GROUP BY key
+INSERT OVERWRITE TABLE e2
+ SELECT key, COUNT(*) GROUP BY key
+PREHOOK: type: QUERY
+POSTHOOK: query: explain FROM src
+INSERT OVERWRITE TABLE e1
+ SELECT key, COUNT(*) WHERE key>450 GROUP BY key
+INSERT OVERWRITE TABLE e2
+ SELECT key, COUNT(*) GROUP BY key
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+ Stage-2 is a root stage
+ Stage-3 depends on stages: Stage-2
+ Stage-0 depends on stages: Stage-3
+ Stage-4 depends on stages: Stage-0
+ Stage-1 depends on stages: Stage-3
+ Stage-5 depends on stages: Stage-1
+
+STAGE PLANS:
+ Stage: Stage-2
+ Spark
+ Edges:
+ Reducer 2 <- Map 1 (GROUP SORT)
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string)
+ outputColumnNames: key
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: key (type: string)
+ sort order: +
+ Map-reduce partition columns: key (type: string)
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Reducer 2
+ Reduce Operator Tree:
+ Forward
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Filter Operator
+ predicate: (KEY._col0 > 450) (type: boolean)
+ Statistics: Num rows: 166 Data size: 1763 Basic stats: COMPLETE Column stats: NONE
+ Group By Operator
+ aggregations: count()
+ keys: KEY._col0 (type: string)
+ mode: complete
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 83 Data size: 881 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: string), UDFToInteger(_col1) (type: int)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 83 Data size: 881 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 83 Data size: 881 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e1
+ Group By Operator
+ aggregations: count()
+ keys: KEY._col0 (type: string)
+ mode: complete
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: string), UDFToInteger(_col1) (type: int)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e2
+
+ Stage: Stage-3
+ Dependency Collection
+
+ Stage: Stage-0
+ Move Operator
+ tables:
+ replace: true
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e1
+
+ Stage: Stage-4
+ Stats-Aggr Operator
+
+ Stage: Stage-1
+ Move Operator
+ tables:
+ replace: true
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e2
+
+ Stage: Stage-5
+ Stats-Aggr Operator
+
+PREHOOK: query: FROM src
+INSERT OVERWRITE TABLE e1
+ SELECT key, COUNT(*) WHERE key>450 GROUP BY key
+INSERT OVERWRITE TABLE e2
+ SELECT key, COUNT(*) GROUP BY key
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@e1
+PREHOOK: Output: default@e2
+[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to
+[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to
+POSTHOOK: query: FROM src
+INSERT OVERWRITE TABLE e1
+ SELECT key, COUNT(*) WHERE key>450 GROUP BY key
+INSERT OVERWRITE TABLE e2
+ SELECT key, COUNT(*) GROUP BY key
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@e1
+POSTHOOK: Output: default@e2
+POSTHOOK: Lineage: e1.count EXPRESSION [(src)src.null, ]
+POSTHOOK: Lineage: e1.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: e2.count EXPRESSION [(src)src.null, ]
+POSTHOOK: Lineage: e2.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+PREHOOK: query: select * from e1
+PREHOOK: type: QUERY
+PREHOOK: Input: default@e1
+#### A masked pattern was here ####
+POSTHOOK: query: select * from e1
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@e1
+#### A masked pattern was here ####
+452 1
+453 1
+454 3
+455 1
+457 1
+458 2
+459 2
+460 1
+462 2
+463 2
+466 3
+467 1
+468 4
+469 5
+470 1
+472 1
+475 1
+477 1
+478 2
+479 1
+480 3
+481 1
+482 1
+483 1
+484 1
+485 1
+487 1
+489 4
+490 1
+491 1
+492 2
+493 1
+494 1
+495 1
+496 1
+497 1
+498 3
+PREHOOK: query: select * from e2
+PREHOOK: type: QUERY
+PREHOOK: Input: default@e2
+#### A masked pattern was here ####
+POSTHOOK: query: select * from e2
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@e2
+#### A masked pattern was here ####
+0 3
+10 1
+100 2
+103 2
+104 2
+105 1
+11 1
+111 1
+113 2
+114 1
+116 1
+118 2
+119 3
+12 2
+120 2
+125 2
+126 1
+128 3
+129 2
+131 1
+133 1
+134 2
+136 1
+137 2
+138 4
+143 1
+145 1
+146 2
+149 2
+15 2
+150 1
+152 2
+153 1
+155 1
+156 1
+157 1
+158 1
+160 1
+162 1
+163 1
+164 2
+165 2
+166 1
+167 3
+168 1
+169 4
+17 1
+170 1
+172 2
+174 2
+175 2
+176 2
+177 1
+178 1
+179 2
+18 2
+180 1
+181 1
+183 1
+186 1
+187 3
+189 1
+19 1
+190 1
+191 2
+192 1
+193 3
+194 1
+195 2
+196 1
+197 2
+199 3
+2 1
+20 1
+200 2
+201 1
+202 1
+203 2
+205 2
+207 2
+208 3
+209 2
+213 2
+214 1
+216 2
+217 2
+218 1
+219 2
+221 2
+222 1
+223 2
+224 2
+226 1
+228 1
+229 2
+230 5
+233 2
+235 1
+237 2
+238 2
+239 2
+24 2
+241 1
+242 2
+244 1
+247 1
+248 1
+249 1
+252 1
+255 2
+256 2
+257 1
+258 1
+26 2
+260 1
+262 1
+263 1
+265 2
+266 1
+27 1
+272 2
+273 3
+274 1
+275 1
+277 4
+278 2
+28 1
+280 2
+281 2
+282 2
+283 1
+284 1
+285 1
+286 1
+287 1
+288 2
+289 1
+291 1
+292 1
+296 1
+298 3
+30 1
+302 1
+305 1
+306 1
+307 2
+308 1
+309 2
+310 1
+311 3
+315 1
+316 3
+317 2
+318 3
+321 2
+322 2
+323 1
+325 2
+327 3
+33 1
+331 2
+332 1
+333 2
+335 1
+336 1
+338 1
+339 1
+34 1
+341 1
+342 2
+344 2
+345 1
+348 5
+35 3
+351 1
+353 2
+356 1
+360 1
+362 1
+364 1
+365 1
+366 1
+367 2
+368 1
+369 3
+37 2
+373 1
+374 1
+375 1
+377 1
+378 1
+379 1
+382 2
+384 3
+386 1
+389 1
+392 1
+393 1
+394 1
+395 2
+396 3
+397 2
+399 2
+4 1
+400 1
+401 5
+402 1
+403 3
+404 2
+406 4
+407 1
+409 3
+41 1
+411 1
+413 2
+414 2
+417 3
+418 1
+419 1
+42 2
+421 1
+424 2
+427 1
+429 2
+43 1
+430 3
+431 3
+432 1
+435 1
+436 1
+437 1
+438 3
+439 2
+44 1
+443 1
+444 1
+446 1
+448 1
+449 1
+452 1
+453 1
+454 3
+455 1
+457 1
+458 2
+459 2
+460 1
+462 2
+463 2
+466 3
+467 1
+468 4
+469 5
+47 1
+470 1
+472 1
+475 1
+477 1
+478 2
+479 1
+480 3
+481 1
+482 1
+483 1
+484 1
+485 1
+487 1
+489 4
+490 1
+491 1
+492 2
+493 1
+494 1
+495 1
+496 1
+497 1
+498 3
+5 3
+51 2
+53 1
+54 1
+57 1
+58 2
+64 1
+65 1
+66 1
+67 2
+69 1
+70 3
+72 2
+74 1
+76 2
+77 1
+78 1
+8 1
+80 1
+82 1
+83 2
+84 2
+85 1
+86 1
+87 1
+9 1
+90 3
+92 1
+95 2
+96 1
+97 2
+98 2
Added: hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby2.q.out?rev=1627210&view=auto
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby2.q.out (added)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby2.q.out Wed Sep 24 03:16:25 2014
@@ -0,0 +1,180 @@
+PREHOOK: query: --HIVE-3699 Multiple insert overwrite into multiple tables query stores same results in all tables
+create table e1 (count int)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@e1
+POSTHOOK: query: --HIVE-3699 Multiple insert overwrite into multiple tables query stores same results in all tables
+create table e1 (count int)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@e1
+PREHOOK: query: create table e2 (percentile double)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@e2
+POSTHOOK: query: create table e2 (percentile double)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@e2
+PREHOOK: query: explain
+FROM (select key, cast(key as double) as value from src order by key) a
+INSERT OVERWRITE TABLE e1
+ SELECT COUNT(*)
+INSERT OVERWRITE TABLE e2
+ SELECT percentile_approx(value, 0.5)
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+FROM (select key, cast(key as double) as value from src order by key) a
+INSERT OVERWRITE TABLE e1
+ SELECT COUNT(*)
+INSERT OVERWRITE TABLE e2
+ SELECT percentile_approx(value, 0.5)
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+ Stage-2 is a root stage
+ Stage-3 depends on stages: Stage-2
+ Stage-0 depends on stages: Stage-3
+ Stage-4 depends on stages: Stage-0
+ Stage-1 depends on stages: Stage-3
+ Stage-5 depends on stages: Stage-1
+
+STAGE PLANS:
+ Stage: Stage-2
+ Spark
+ Edges:
+ Reducer 2 <- Map 1 (GROUP SORT)
+ Reducer 3 <- Reducer 2 (GROUP)
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string), UDFToDouble(key) (type: double)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: string)
+ sort order: +
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col1 (type: double)
+ Reducer 2
+ Reduce Operator Tree:
+ Select Operator
+ expressions: VALUE._col0 (type: double)
+ outputColumnNames: _col1
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ sort order:
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col1 (type: double)
+ Reducer 3
+ Reduce Operator Tree:
+ Forward
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Group By Operator
+ aggregations: count()
+ mode: complete
+ outputColumnNames: _col0
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: UDFToInteger(_col0) (type: int)
+ outputColumnNames: _col0
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e1
+ Group By Operator
+ aggregations: percentile_approx(VALUE._col0, 0.5)
+ mode: complete
+ outputColumnNames: _col0
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: double)
+ outputColumnNames: _col0
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e2
+
+ Stage: Stage-3
+ Dependency Collection
+
+ Stage: Stage-0
+ Move Operator
+ tables:
+ replace: true
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e1
+
+ Stage: Stage-4
+ Stats-Aggr Operator
+
+ Stage: Stage-1
+ Move Operator
+ tables:
+ replace: true
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e2
+
+ Stage: Stage-5
+ Stats-Aggr Operator
+
+PREHOOK: query: FROM (select key, cast(key as double) as value from src order by key) a
+INSERT OVERWRITE TABLE e1
+ SELECT COUNT(*)
+INSERT OVERWRITE TABLE e2
+ SELECT percentile_approx(value, 0.5)
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@e1
+PREHOOK: Output: default@e2
+[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to
+[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to
+POSTHOOK: query: FROM (select key, cast(key as double) as value from src order by key) a
+INSERT OVERWRITE TABLE e1
+ SELECT COUNT(*)
+INSERT OVERWRITE TABLE e2
+ SELECT percentile_approx(value, 0.5)
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@e1
+POSTHOOK: Output: default@e2
+POSTHOOK: Lineage: e1.count EXPRESSION [(src)src.null, ]
+POSTHOOK: Lineage: e2.percentile EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+PREHOOK: query: select * from e1
+PREHOOK: type: QUERY
+PREHOOK: Input: default@e1
+#### A masked pattern was here ####
+POSTHOOK: query: select * from e1
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@e1
+#### A masked pattern was here ####
+500
+PREHOOK: query: select * from e2
+PREHOOK: type: QUERY
+PREHOOK: Input: default@e2
+#### A masked pattern was here ####
+POSTHOOK: query: select * from e2
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@e2
+#### A masked pattern was here ####
+255.5
Added: hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby3.q.out?rev=1627210&view=auto
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby3.q.out (added)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby3.q.out Wed Sep 24 03:16:25 2014
@@ -0,0 +1,1951 @@
+PREHOOK: query: create table e1 (key string, keyD double)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@e1
+POSTHOOK: query: create table e1 (key string, keyD double)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@e1
+PREHOOK: query: create table e2 (key string, keyD double, value string)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@e2
+POSTHOOK: query: create table e2 (key string, keyD double, value string)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@e2
+PREHOOK: query: create table e3 (key string, keyD double)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@e3
+POSTHOOK: query: create table e3 (key string, keyD double)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@e3
+PREHOOK: query: explain
+FROM (select key, cast(key as double) as keyD, value from src order by key) a
+INSERT OVERWRITE TABLE e1
+ SELECT key, COUNT(distinct value) group by key
+INSERT OVERWRITE TABLE e2
+ SELECT key, sum(keyD), value group by key, value
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+FROM (select key, cast(key as double) as keyD, value from src order by key) a
+INSERT OVERWRITE TABLE e1
+ SELECT key, COUNT(distinct value) group by key
+INSERT OVERWRITE TABLE e2
+ SELECT key, sum(keyD), value group by key, value
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+ Stage-2 is a root stage
+ Stage-3 depends on stages: Stage-2
+ Stage-0 depends on stages: Stage-3
+ Stage-4 depends on stages: Stage-0
+ Stage-1 depends on stages: Stage-3
+ Stage-5 depends on stages: Stage-1
+
+STAGE PLANS:
+ Stage: Stage-2
+ Spark
+ Edges:
+ Reducer 2 <- Map 1 (GROUP SORT)
+ Reducer 3 <- Reducer 2 (GROUP SORT)
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string), UDFToDouble(key) (type: double), value (type: string)
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: string)
+ sort order: +
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col1 (type: double), _col2 (type: string)
+ Reducer 2
+ Reduce Operator Tree:
+ Select Operator
+ expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: double), VALUE._col1 (type: string)
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: string), _col2 (type: string)
+ sort order: ++
+ Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col1 (type: double)
+ Reducer 3
+ Reduce Operator Tree:
+ Forward
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Group By Operator
+ aggregations: count(DISTINCT KEY._col1:0._col0)
+ keys: KEY._col0 (type: string)
+ mode: complete
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: string), UDFToDouble(_col1) (type: double)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e1
+ Group By Operator
+ aggregations: sum(VALUE._col0)
+ keys: KEY._col0 (type: string), KEY._col1:0._col0 (type: string)
+ mode: complete
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: string), _col2 (type: double), _col1 (type: string)
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e2
+
+ Stage: Stage-3
+ Dependency Collection
+
+ Stage: Stage-0
+ Move Operator
+ tables:
+ replace: true
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e1
+
+ Stage: Stage-4
+ Stats-Aggr Operator
+
+ Stage: Stage-1
+ Move Operator
+ tables:
+ replace: true
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e2
+
+ Stage: Stage-5
+ Stats-Aggr Operator
+
+PREHOOK: query: explain
+FROM (select key, cast(key as double) as keyD, value from src order by key) a
+INSERT OVERWRITE TABLE e2
+ SELECT key, sum(keyD), value group by key, value
+INSERT OVERWRITE TABLE e1
+ SELECT key, COUNT(distinct value) group by key
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+FROM (select key, cast(key as double) as keyD, value from src order by key) a
+INSERT OVERWRITE TABLE e2
+ SELECT key, sum(keyD), value group by key, value
+INSERT OVERWRITE TABLE e1
+ SELECT key, COUNT(distinct value) group by key
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+ Stage-2 is a root stage
+ Stage-3 depends on stages: Stage-2
+ Stage-0 depends on stages: Stage-3
+ Stage-4 depends on stages: Stage-0
+ Stage-1 depends on stages: Stage-3
+ Stage-5 depends on stages: Stage-1
+
+STAGE PLANS:
+ Stage: Stage-2
+ Spark
+ Edges:
+ Reducer 2 <- Map 1 (GROUP SORT)
+ Reducer 3 <- Reducer 2 (GROUP SORT)
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string), UDFToDouble(key) (type: double), value (type: string)
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: string)
+ sort order: +
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col1 (type: double), _col2 (type: string)
+ Reducer 2
+ Reduce Operator Tree:
+ Select Operator
+ expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: double), VALUE._col1 (type: string)
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: string), _col2 (type: string)
+ sort order: ++
+ Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col1 (type: double)
+ Reducer 3
+ Reduce Operator Tree:
+ Forward
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Group By Operator
+ aggregations: count(DISTINCT KEY._col1:0._col0)
+ keys: KEY._col0 (type: string)
+ mode: complete
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: string), UDFToDouble(_col1) (type: double)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e1
+ Group By Operator
+ aggregations: sum(VALUE._col0)
+ keys: KEY._col0 (type: string), KEY._col1:0._col0 (type: string)
+ mode: complete
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: string), _col2 (type: double), _col1 (type: string)
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e2
+
+ Stage: Stage-3
+ Dependency Collection
+
+ Stage: Stage-0
+ Move Operator
+ tables:
+ replace: true
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e1
+
+ Stage: Stage-4
+ Stats-Aggr Operator
+
+ Stage: Stage-1
+ Move Operator
+ tables:
+ replace: true
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e2
+
+ Stage: Stage-5
+ Stats-Aggr Operator
+
+PREHOOK: query: FROM (select key, cast(key as double) as keyD, value from src order by key) a
+INSERT OVERWRITE TABLE e1
+ SELECT key, COUNT(distinct value) group by key
+INSERT OVERWRITE TABLE e2
+ SELECT key, sum(keyD), value group by key, value
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@e1
+PREHOOK: Output: default@e2
+[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to
+[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to
+POSTHOOK: query: FROM (select key, cast(key as double) as keyD, value from src order by key) a
+INSERT OVERWRITE TABLE e1
+ SELECT key, COUNT(distinct value) group by key
+INSERT OVERWRITE TABLE e2
+ SELECT key, sum(keyD), value group by key, value
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@e1
+POSTHOOK: Output: default@e2
+POSTHOOK: Lineage: e1.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: e1.keyd EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: e2.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: e2.keyd EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: e2.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
+PREHOOK: query: select * from e1
+PREHOOK: type: QUERY
+PREHOOK: Input: default@e1
+#### A masked pattern was here ####
+POSTHOOK: query: select * from e1
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@e1
+#### A masked pattern was here ####
+0 1.0
+10 1.0
+100 1.0
+103 1.0
+104 1.0
+105 1.0
+11 1.0
+111 1.0
+113 1.0
+114 1.0
+116 1.0
+118 1.0
+119 1.0
+12 1.0
+120 1.0
+125 1.0
+126 1.0
+128 1.0
+129 1.0
+131 1.0
+133 1.0
+134 1.0
+136 1.0
+137 1.0
+138 1.0
+143 1.0
+145 1.0
+146 1.0
+149 1.0
+15 1.0
+150 1.0
+152 1.0
+153 1.0
+155 1.0
+156 1.0
+157 1.0
+158 1.0
+160 1.0
+162 1.0
+163 1.0
+164 1.0
+165 1.0
+166 1.0
+167 1.0
+168 1.0
+169 1.0
+17 1.0
+170 1.0
+172 1.0
+174 1.0
+175 1.0
+176 1.0
+177 1.0
+178 1.0
+179 1.0
+18 1.0
+180 1.0
+181 1.0
+183 1.0
+186 1.0
+187 1.0
+189 1.0
+19 1.0
+190 1.0
+191 1.0
+192 1.0
+193 1.0
+194 1.0
+195 1.0
+196 1.0
+197 1.0
+199 1.0
+2 1.0
+20 1.0
+200 1.0
+201 1.0
+202 1.0
+203 1.0
+205 1.0
+207 1.0
+208 1.0
+209 1.0
+213 1.0
+214 1.0
+216 1.0
+217 1.0
+218 1.0
+219 1.0
+221 1.0
+222 1.0
+223 1.0
+224 1.0
+226 1.0
+228 1.0
+229 1.0
+230 1.0
+233 1.0
+235 1.0
+237 1.0
+238 1.0
+239 1.0
+24 1.0
+241 1.0
+242 1.0
+244 1.0
+247 1.0
+248 1.0
+249 1.0
+252 1.0
+255 1.0
+256 1.0
+257 1.0
+258 1.0
+26 1.0
+260 1.0
+262 1.0
+263 1.0
+265 1.0
+266 1.0
+27 1.0
+272 1.0
+273 1.0
+274 1.0
+275 1.0
+277 1.0
+278 1.0
+28 1.0
+280 1.0
+281 1.0
+282 1.0
+283 1.0
+284 1.0
+285 1.0
+286 1.0
+287 1.0
+288 1.0
+289 1.0
+291 1.0
+292 1.0
+296 1.0
+298 1.0
+30 1.0
+302 1.0
+305 1.0
+306 1.0
+307 1.0
+308 1.0
+309 1.0
+310 1.0
+311 1.0
+315 1.0
+316 1.0
+317 1.0
+318 1.0
+321 1.0
+322 1.0
+323 1.0
+325 1.0
+327 1.0
+33 1.0
+331 1.0
+332 1.0
+333 1.0
+335 1.0
+336 1.0
+338 1.0
+339 1.0
+34 1.0
+341 1.0
+342 1.0
+344 1.0
+345 1.0
+348 1.0
+35 1.0
+351 1.0
+353 1.0
+356 1.0
+360 1.0
+362 1.0
+364 1.0
+365 1.0
+366 1.0
+367 1.0
+368 1.0
+369 1.0
+37 1.0
+373 1.0
+374 1.0
+375 1.0
+377 1.0
+378 1.0
+379 1.0
+382 1.0
+384 1.0
+386 1.0
+389 1.0
+392 1.0
+393 1.0
+394 1.0
+395 1.0
+396 1.0
+397 1.0
+399 1.0
+4 1.0
+400 1.0
+401 1.0
+402 1.0
+403 1.0
+404 1.0
+406 1.0
+407 1.0
+409 1.0
+41 1.0
+411 1.0
+413 1.0
+414 1.0
+417 1.0
+418 1.0
+419 1.0
+42 1.0
+421 1.0
+424 1.0
+427 1.0
+429 1.0
+43 1.0
+430 1.0
+431 1.0
+432 1.0
+435 1.0
+436 1.0
+437 1.0
+438 1.0
+439 1.0
+44 1.0
+443 1.0
+444 1.0
+446 1.0
+448 1.0
+449 1.0
+452 1.0
+453 1.0
+454 1.0
+455 1.0
+457 1.0
+458 1.0
+459 1.0
+460 1.0
+462 1.0
+463 1.0
+466 1.0
+467 1.0
+468 1.0
+469 1.0
+47 1.0
+470 1.0
+472 1.0
+475 1.0
+477 1.0
+478 1.0
+479 1.0
+480 1.0
+481 1.0
+482 1.0
+483 1.0
+484 1.0
+485 1.0
+487 1.0
+489 1.0
+490 1.0
+491 1.0
+492 1.0
+493 1.0
+494 1.0
+495 1.0
+496 1.0
+497 1.0
+498 1.0
+5 1.0
+51 1.0
+53 1.0
+54 1.0
+57 1.0
+58 1.0
+64 1.0
+65 1.0
+66 1.0
+67 1.0
+69 1.0
+70 1.0
+72 1.0
+74 1.0
+76 1.0
+77 1.0
+78 1.0
+8 1.0
+80 1.0
+82 1.0
+83 1.0
+84 1.0
+85 1.0
+86 1.0
+87 1.0
+9 1.0
+90 1.0
+92 1.0
+95 1.0
+96 1.0
+97 1.0
+98 1.0
+PREHOOK: query: select * from e2
+PREHOOK: type: QUERY
+PREHOOK: Input: default@e2
+#### A masked pattern was here ####
+POSTHOOK: query: select * from e2
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@e2
+#### A masked pattern was here ####
+0 0.0 val_0
+10 10.0 val_10
+100 200.0 val_100
+103 206.0 val_103
+104 208.0 val_104
+105 105.0 val_105
+11 11.0 val_11
+111 111.0 val_111
+113 226.0 val_113
+114 114.0 val_114
+116 116.0 val_116
+118 236.0 val_118
+119 357.0 val_119
+12 24.0 val_12
+120 240.0 val_120
+125 250.0 val_125
+126 126.0 val_126
+128 384.0 val_128
+129 258.0 val_129
+131 131.0 val_131
+133 133.0 val_133
+134 268.0 val_134
+136 136.0 val_136
+137 274.0 val_137
+138 552.0 val_138
+143 143.0 val_143
+145 145.0 val_145
+146 292.0 val_146
+149 298.0 val_149
+15 30.0 val_15
+150 150.0 val_150
+152 304.0 val_152
+153 153.0 val_153
+155 155.0 val_155
+156 156.0 val_156
+157 157.0 val_157
+158 158.0 val_158
+160 160.0 val_160
+162 162.0 val_162
+163 163.0 val_163
+164 328.0 val_164
+165 330.0 val_165
+166 166.0 val_166
+167 501.0 val_167
+168 168.0 val_168
+169 676.0 val_169
+17 17.0 val_17
+170 170.0 val_170
+172 344.0 val_172
+174 348.0 val_174
+175 350.0 val_175
+176 352.0 val_176
+177 177.0 val_177
+178 178.0 val_178
+179 358.0 val_179
+18 36.0 val_18
+180 180.0 val_180
+181 181.0 val_181
+183 183.0 val_183
+186 186.0 val_186
+187 561.0 val_187
+189 189.0 val_189
+19 19.0 val_19
+190 190.0 val_190
+191 382.0 val_191
+192 192.0 val_192
+193 579.0 val_193
+194 194.0 val_194
+195 390.0 val_195
+196 196.0 val_196
+197 394.0 val_197
+199 597.0 val_199
+2 2.0 val_2
+20 20.0 val_20
+200 400.0 val_200
+201 201.0 val_201
+202 202.0 val_202
+203 406.0 val_203
+205 410.0 val_205
+207 414.0 val_207
+208 624.0 val_208
+209 418.0 val_209
+213 426.0 val_213
+214 214.0 val_214
+216 432.0 val_216
+217 434.0 val_217
+218 218.0 val_218
+219 438.0 val_219
+221 442.0 val_221
+222 222.0 val_222
+223 446.0 val_223
+224 448.0 val_224
+226 226.0 val_226
+228 228.0 val_228
+229 458.0 val_229
+230 1150.0 val_230
+233 466.0 val_233
+235 235.0 val_235
+237 474.0 val_237
+238 476.0 val_238
+239 478.0 val_239
+24 48.0 val_24
+241 241.0 val_241
+242 484.0 val_242
+244 244.0 val_244
+247 247.0 val_247
+248 248.0 val_248
+249 249.0 val_249
+252 252.0 val_252
+255 510.0 val_255
+256 512.0 val_256
+257 257.0 val_257
+258 258.0 val_258
+26 52.0 val_26
+260 260.0 val_260
+262 262.0 val_262
+263 263.0 val_263
+265 530.0 val_265
+266 266.0 val_266
+27 27.0 val_27
+272 544.0 val_272
+273 819.0 val_273
+274 274.0 val_274
+275 275.0 val_275
+277 1108.0 val_277
+278 556.0 val_278
+28 28.0 val_28
+280 560.0 val_280
+281 562.0 val_281
+282 564.0 val_282
+283 283.0 val_283
+284 284.0 val_284
+285 285.0 val_285
+286 286.0 val_286
+287 287.0 val_287
+288 576.0 val_288
+289 289.0 val_289
+291 291.0 val_291
+292 292.0 val_292
+296 296.0 val_296
+298 894.0 val_298
+30 30.0 val_30
+302 302.0 val_302
+305 305.0 val_305
+306 306.0 val_306
+307 614.0 val_307
+308 308.0 val_308
+309 618.0 val_309
+310 310.0 val_310
+311 933.0 val_311
+315 315.0 val_315
+316 948.0 val_316
+317 634.0 val_317
+318 954.0 val_318
+321 642.0 val_321
+322 644.0 val_322
+323 323.0 val_323
+325 650.0 val_325
+327 981.0 val_327
+33 33.0 val_33
+331 662.0 val_331
+332 332.0 val_332
+333 666.0 val_333
+335 335.0 val_335
+336 336.0 val_336
+338 338.0 val_338
+339 339.0 val_339
+34 34.0 val_34
+341 341.0 val_341
+342 684.0 val_342
+344 688.0 val_344
+345 345.0 val_345
+348 1740.0 val_348
+35 105.0 val_35
+351 351.0 val_351
+353 706.0 val_353
+356 356.0 val_356
+360 360.0 val_360
+362 362.0 val_362
+364 364.0 val_364
+365 365.0 val_365
+366 366.0 val_366
+367 734.0 val_367
+368 368.0 val_368
+369 1107.0 val_369
+37 74.0 val_37
+373 373.0 val_373
+374 374.0 val_374
+375 375.0 val_375
+377 377.0 val_377
+378 378.0 val_378
+379 379.0 val_379
+382 764.0 val_382
+384 1152.0 val_384
+386 386.0 val_386
+389 389.0 val_389
+392 392.0 val_392
+393 393.0 val_393
+394 394.0 val_394
+395 790.0 val_395
+396 1188.0 val_396
+397 794.0 val_397
+399 798.0 val_399
+4 4.0 val_4
+400 400.0 val_400
+401 2005.0 val_401
+402 402.0 val_402
+403 1209.0 val_403
+404 808.0 val_404
+406 1624.0 val_406
+407 407.0 val_407
+409 1227.0 val_409
+41 41.0 val_41
+411 411.0 val_411
+413 826.0 val_413
+414 828.0 val_414
+417 1251.0 val_417
+418 418.0 val_418
+419 419.0 val_419
+42 84.0 val_42
+421 421.0 val_421
+424 848.0 val_424
+427 427.0 val_427
+429 858.0 val_429
+43 43.0 val_43
+430 1290.0 val_430
+431 1293.0 val_431
+432 432.0 val_432
+435 435.0 val_435
+436 436.0 val_436
+437 437.0 val_437
+438 1314.0 val_438
+439 878.0 val_439
+44 44.0 val_44
+443 443.0 val_443
+444 444.0 val_444
+446 446.0 val_446
+448 448.0 val_448
+449 449.0 val_449
+452 452.0 val_452
+453 453.0 val_453
+454 1362.0 val_454
+455 455.0 val_455
+457 457.0 val_457
+458 916.0 val_458
+459 918.0 val_459
+460 460.0 val_460
+462 924.0 val_462
+463 926.0 val_463
+466 1398.0 val_466
+467 467.0 val_467
+468 1872.0 val_468
+469 2345.0 val_469
+47 47.0 val_47
+470 470.0 val_470
+472 472.0 val_472
+475 475.0 val_475
+477 477.0 val_477
+478 956.0 val_478
+479 479.0 val_479
+480 1440.0 val_480
+481 481.0 val_481
+482 482.0 val_482
+483 483.0 val_483
+484 484.0 val_484
+485 485.0 val_485
+487 487.0 val_487
+489 1956.0 val_489
+490 490.0 val_490
+491 491.0 val_491
+492 984.0 val_492
+493 493.0 val_493
+494 494.0 val_494
+495 495.0 val_495
+496 496.0 val_496
+497 497.0 val_497
+498 1494.0 val_498
+5 15.0 val_5
+51 102.0 val_51
+53 53.0 val_53
+54 54.0 val_54
+57 57.0 val_57
+58 116.0 val_58
+64 64.0 val_64
+65 65.0 val_65
+66 66.0 val_66
+67 134.0 val_67
+69 69.0 val_69
+70 210.0 val_70
+72 144.0 val_72
+74 74.0 val_74
+76 152.0 val_76
+77 77.0 val_77
+78 78.0 val_78
+8 8.0 val_8
+80 80.0 val_80
+82 82.0 val_82
+83 166.0 val_83
+84 168.0 val_84
+85 85.0 val_85
+86 86.0 val_86
+87 87.0 val_87
+9 9.0 val_9
+90 270.0 val_90
+92 92.0 val_92
+95 190.0 val_95
+96 96.0 val_96
+97 194.0 val_97
+98 196.0 val_98
+PREHOOK: query: FROM (select key, cast(key as double) as keyD, value from src order by key) a
+INSERT OVERWRITE TABLE e2
+ SELECT key, sum(keyD), value group by key, value
+INSERT OVERWRITE TABLE e1
+ SELECT key, COUNT(distinct value) group by key
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@e1
+PREHOOK: Output: default@e2
+[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to
+[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to
+POSTHOOK: query: FROM (select key, cast(key as double) as keyD, value from src order by key) a
+INSERT OVERWRITE TABLE e2
+ SELECT key, sum(keyD), value group by key, value
+INSERT OVERWRITE TABLE e1
+ SELECT key, COUNT(distinct value) group by key
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@e1
+POSTHOOK: Output: default@e2
+POSTHOOK: Lineage: e1.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: e1.keyd EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: e2.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: e2.keyd EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: e2.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
+PREHOOK: query: select * from e1
+PREHOOK: type: QUERY
+PREHOOK: Input: default@e1
+#### A masked pattern was here ####
+POSTHOOK: query: select * from e1
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@e1
+#### A masked pattern was here ####
+0 1.0
+10 1.0
+100 1.0
+103 1.0
+104 1.0
+105 1.0
+11 1.0
+111 1.0
+113 1.0
+114 1.0
+116 1.0
+118 1.0
+119 1.0
+12 1.0
+120 1.0
+125 1.0
+126 1.0
+128 1.0
+129 1.0
+131 1.0
+133 1.0
+134 1.0
+136 1.0
+137 1.0
+138 1.0
+143 1.0
+145 1.0
+146 1.0
+149 1.0
+15 1.0
+150 1.0
+152 1.0
+153 1.0
+155 1.0
+156 1.0
+157 1.0
+158 1.0
+160 1.0
+162 1.0
+163 1.0
+164 1.0
+165 1.0
+166 1.0
+167 1.0
+168 1.0
+169 1.0
+17 1.0
+170 1.0
+172 1.0
+174 1.0
+175 1.0
+176 1.0
+177 1.0
+178 1.0
+179 1.0
+18 1.0
+180 1.0
+181 1.0
+183 1.0
+186 1.0
+187 1.0
+189 1.0
+19 1.0
+190 1.0
+191 1.0
+192 1.0
+193 1.0
+194 1.0
+195 1.0
+196 1.0
+197 1.0
+199 1.0
+2 1.0
+20 1.0
+200 1.0
+201 1.0
+202 1.0
+203 1.0
+205 1.0
+207 1.0
+208 1.0
+209 1.0
+213 1.0
+214 1.0
+216 1.0
+217 1.0
+218 1.0
+219 1.0
+221 1.0
+222 1.0
+223 1.0
+224 1.0
+226 1.0
+228 1.0
+229 1.0
+230 1.0
+233 1.0
+235 1.0
+237 1.0
+238 1.0
+239 1.0
+24 1.0
+241 1.0
+242 1.0
+244 1.0
+247 1.0
+248 1.0
+249 1.0
+252 1.0
+255 1.0
+256 1.0
+257 1.0
+258 1.0
+26 1.0
+260 1.0
+262 1.0
+263 1.0
+265 1.0
+266 1.0
+27 1.0
+272 1.0
+273 1.0
+274 1.0
+275 1.0
+277 1.0
+278 1.0
+28 1.0
+280 1.0
+281 1.0
+282 1.0
+283 1.0
+284 1.0
+285 1.0
+286 1.0
+287 1.0
+288 1.0
+289 1.0
+291 1.0
+292 1.0
+296 1.0
+298 1.0
+30 1.0
+302 1.0
+305 1.0
+306 1.0
+307 1.0
+308 1.0
+309 1.0
+310 1.0
+311 1.0
+315 1.0
+316 1.0
+317 1.0
+318 1.0
+321 1.0
+322 1.0
+323 1.0
+325 1.0
+327 1.0
+33 1.0
+331 1.0
+332 1.0
+333 1.0
+335 1.0
+336 1.0
+338 1.0
+339 1.0
+34 1.0
+341 1.0
+342 1.0
+344 1.0
+345 1.0
+348 1.0
+35 1.0
+351 1.0
+353 1.0
+356 1.0
+360 1.0
+362 1.0
+364 1.0
+365 1.0
+366 1.0
+367 1.0
+368 1.0
+369 1.0
+37 1.0
+373 1.0
+374 1.0
+375 1.0
+377 1.0
+378 1.0
+379 1.0
+382 1.0
+384 1.0
+386 1.0
+389 1.0
+392 1.0
+393 1.0
+394 1.0
+395 1.0
+396 1.0
+397 1.0
+399 1.0
+4 1.0
+400 1.0
+401 1.0
+402 1.0
+403 1.0
+404 1.0
+406 1.0
+407 1.0
+409 1.0
+41 1.0
+411 1.0
+413 1.0
+414 1.0
+417 1.0
+418 1.0
+419 1.0
+42 1.0
+421 1.0
+424 1.0
+427 1.0
+429 1.0
+43 1.0
+430 1.0
+431 1.0
+432 1.0
+435 1.0
+436 1.0
+437 1.0
+438 1.0
+439 1.0
+44 1.0
+443 1.0
+444 1.0
+446 1.0
+448 1.0
+449 1.0
+452 1.0
+453 1.0
+454 1.0
+455 1.0
+457 1.0
+458 1.0
+459 1.0
+460 1.0
+462 1.0
+463 1.0
+466 1.0
+467 1.0
+468 1.0
+469 1.0
+47 1.0
+470 1.0
+472 1.0
+475 1.0
+477 1.0
+478 1.0
+479 1.0
+480 1.0
+481 1.0
+482 1.0
+483 1.0
+484 1.0
+485 1.0
+487 1.0
+489 1.0
+490 1.0
+491 1.0
+492 1.0
+493 1.0
+494 1.0
+495 1.0
+496 1.0
+497 1.0
+498 1.0
+5 1.0
+51 1.0
+53 1.0
+54 1.0
+57 1.0
+58 1.0
+64 1.0
+65 1.0
+66 1.0
+67 1.0
+69 1.0
+70 1.0
+72 1.0
+74 1.0
+76 1.0
+77 1.0
+78 1.0
+8 1.0
+80 1.0
+82 1.0
+83 1.0
+84 1.0
+85 1.0
+86 1.0
+87 1.0
+9 1.0
+90 1.0
+92 1.0
+95 1.0
+96 1.0
+97 1.0
+98 1.0
+PREHOOK: query: select * from e2
+PREHOOK: type: QUERY
+PREHOOK: Input: default@e2
+#### A masked pattern was here ####
+POSTHOOK: query: select * from e2
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@e2
+#### A masked pattern was here ####
+0 0.0 val_0
+10 10.0 val_10
+100 200.0 val_100
+103 206.0 val_103
+104 208.0 val_104
+105 105.0 val_105
+11 11.0 val_11
+111 111.0 val_111
+113 226.0 val_113
+114 114.0 val_114
+116 116.0 val_116
+118 236.0 val_118
+119 357.0 val_119
+12 24.0 val_12
+120 240.0 val_120
+125 250.0 val_125
+126 126.0 val_126
+128 384.0 val_128
+129 258.0 val_129
+131 131.0 val_131
+133 133.0 val_133
+134 268.0 val_134
+136 136.0 val_136
+137 274.0 val_137
+138 552.0 val_138
+143 143.0 val_143
+145 145.0 val_145
+146 292.0 val_146
+149 298.0 val_149
+15 30.0 val_15
+150 150.0 val_150
+152 304.0 val_152
+153 153.0 val_153
+155 155.0 val_155
+156 156.0 val_156
+157 157.0 val_157
+158 158.0 val_158
+160 160.0 val_160
+162 162.0 val_162
+163 163.0 val_163
+164 328.0 val_164
+165 330.0 val_165
+166 166.0 val_166
+167 501.0 val_167
+168 168.0 val_168
+169 676.0 val_169
+17 17.0 val_17
+170 170.0 val_170
+172 344.0 val_172
+174 348.0 val_174
+175 350.0 val_175
+176 352.0 val_176
+177 177.0 val_177
+178 178.0 val_178
+179 358.0 val_179
+18 36.0 val_18
+180 180.0 val_180
+181 181.0 val_181
+183 183.0 val_183
+186 186.0 val_186
+187 561.0 val_187
+189 189.0 val_189
+19 19.0 val_19
+190 190.0 val_190
+191 382.0 val_191
+192 192.0 val_192
+193 579.0 val_193
+194 194.0 val_194
+195 390.0 val_195
+196 196.0 val_196
+197 394.0 val_197
+199 597.0 val_199
+2 2.0 val_2
+20 20.0 val_20
+200 400.0 val_200
+201 201.0 val_201
+202 202.0 val_202
+203 406.0 val_203
+205 410.0 val_205
+207 414.0 val_207
+208 624.0 val_208
+209 418.0 val_209
+213 426.0 val_213
+214 214.0 val_214
+216 432.0 val_216
+217 434.0 val_217
+218 218.0 val_218
+219 438.0 val_219
+221 442.0 val_221
+222 222.0 val_222
+223 446.0 val_223
+224 448.0 val_224
+226 226.0 val_226
+228 228.0 val_228
+229 458.0 val_229
+230 1150.0 val_230
+233 466.0 val_233
+235 235.0 val_235
+237 474.0 val_237
+238 476.0 val_238
+239 478.0 val_239
+24 48.0 val_24
+241 241.0 val_241
+242 484.0 val_242
+244 244.0 val_244
+247 247.0 val_247
+248 248.0 val_248
+249 249.0 val_249
+252 252.0 val_252
+255 510.0 val_255
+256 512.0 val_256
+257 257.0 val_257
+258 258.0 val_258
+26 52.0 val_26
+260 260.0 val_260
+262 262.0 val_262
+263 263.0 val_263
+265 530.0 val_265
+266 266.0 val_266
+27 27.0 val_27
+272 544.0 val_272
+273 819.0 val_273
+274 274.0 val_274
+275 275.0 val_275
+277 1108.0 val_277
+278 556.0 val_278
+28 28.0 val_28
+280 560.0 val_280
+281 562.0 val_281
+282 564.0 val_282
+283 283.0 val_283
+284 284.0 val_284
+285 285.0 val_285
+286 286.0 val_286
+287 287.0 val_287
+288 576.0 val_288
+289 289.0 val_289
+291 291.0 val_291
+292 292.0 val_292
+296 296.0 val_296
+298 894.0 val_298
+30 30.0 val_30
+302 302.0 val_302
+305 305.0 val_305
+306 306.0 val_306
+307 614.0 val_307
+308 308.0 val_308
+309 618.0 val_309
+310 310.0 val_310
+311 933.0 val_311
+315 315.0 val_315
+316 948.0 val_316
+317 634.0 val_317
+318 954.0 val_318
+321 642.0 val_321
+322 644.0 val_322
+323 323.0 val_323
+325 650.0 val_325
+327 981.0 val_327
+33 33.0 val_33
+331 662.0 val_331
+332 332.0 val_332
+333 666.0 val_333
+335 335.0 val_335
+336 336.0 val_336
+338 338.0 val_338
+339 339.0 val_339
+34 34.0 val_34
+341 341.0 val_341
+342 684.0 val_342
+344 688.0 val_344
+345 345.0 val_345
+348 1740.0 val_348
+35 105.0 val_35
+351 351.0 val_351
+353 706.0 val_353
+356 356.0 val_356
+360 360.0 val_360
+362 362.0 val_362
+364 364.0 val_364
+365 365.0 val_365
+366 366.0 val_366
+367 734.0 val_367
+368 368.0 val_368
+369 1107.0 val_369
+37 74.0 val_37
+373 373.0 val_373
+374 374.0 val_374
+375 375.0 val_375
+377 377.0 val_377
+378 378.0 val_378
+379 379.0 val_379
+382 764.0 val_382
+384 1152.0 val_384
+386 386.0 val_386
+389 389.0 val_389
+392 392.0 val_392
+393 393.0 val_393
+394 394.0 val_394
+395 790.0 val_395
+396 1188.0 val_396
+397 794.0 val_397
+399 798.0 val_399
+4 4.0 val_4
+400 400.0 val_400
+401 2005.0 val_401
+402 402.0 val_402
+403 1209.0 val_403
+404 808.0 val_404
+406 1624.0 val_406
+407 407.0 val_407
+409 1227.0 val_409
+41 41.0 val_41
+411 411.0 val_411
+413 826.0 val_413
+414 828.0 val_414
+417 1251.0 val_417
+418 418.0 val_418
+419 419.0 val_419
+42 84.0 val_42
+421 421.0 val_421
+424 848.0 val_424
+427 427.0 val_427
+429 858.0 val_429
+43 43.0 val_43
+430 1290.0 val_430
+431 1293.0 val_431
+432 432.0 val_432
+435 435.0 val_435
+436 436.0 val_436
+437 437.0 val_437
+438 1314.0 val_438
+439 878.0 val_439
+44 44.0 val_44
+443 443.0 val_443
+444 444.0 val_444
+446 446.0 val_446
+448 448.0 val_448
+449 449.0 val_449
+452 452.0 val_452
+453 453.0 val_453
+454 1362.0 val_454
+455 455.0 val_455
+457 457.0 val_457
+458 916.0 val_458
+459 918.0 val_459
+460 460.0 val_460
+462 924.0 val_462
+463 926.0 val_463
+466 1398.0 val_466
+467 467.0 val_467
+468 1872.0 val_468
+469 2345.0 val_469
+47 47.0 val_47
+470 470.0 val_470
+472 472.0 val_472
+475 475.0 val_475
+477 477.0 val_477
+478 956.0 val_478
+479 479.0 val_479
+480 1440.0 val_480
+481 481.0 val_481
+482 482.0 val_482
+483 483.0 val_483
+484 484.0 val_484
+485 485.0 val_485
+487 487.0 val_487
+489 1956.0 val_489
+490 490.0 val_490
+491 491.0 val_491
+492 984.0 val_492
+493 493.0 val_493
+494 494.0 val_494
+495 495.0 val_495
+496 496.0 val_496
+497 497.0 val_497
+498 1494.0 val_498
+5 15.0 val_5
+51 102.0 val_51
+53 53.0 val_53
+54 54.0 val_54
+57 57.0 val_57
+58 116.0 val_58
+64 64.0 val_64
+65 65.0 val_65
+66 66.0 val_66
+67 134.0 val_67
+69 69.0 val_69
+70 210.0 val_70
+72 144.0 val_72
+74 74.0 val_74
+76 152.0 val_76
+77 77.0 val_77
+78 78.0 val_78
+8 8.0 val_8
+80 80.0 val_80
+82 82.0 val_82
+83 166.0 val_83
+84 168.0 val_84
+85 85.0 val_85
+86 86.0 val_86
+87 87.0 val_87
+9 9.0 val_9
+90 270.0 val_90
+92 92.0 val_92
+95 190.0 val_95
+96 96.0 val_96
+97 194.0 val_97
+98 196.0 val_98
+PREHOOK: query: explain
+from src
+insert overwrite table e1
+select key, count(distinct value) group by key
+insert overwrite table e3
+select value, count(distinct key) group by value
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+from src
+insert overwrite table e1
+select key, count(distinct value) group by key
+insert overwrite table e3
+select value, count(distinct key) group by value
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+ Stage-2 is a root stage
+ Stage-4 depends on stages: Stage-2
+ Stage-3 depends on stages: Stage-4, Stage-5
+ Stage-0 depends on stages: Stage-3
+ Stage-6 depends on stages: Stage-0
+ Stage-1 depends on stages: Stage-3
+ Stage-7 depends on stages: Stage-1
+ Stage-5 depends on stages: Stage-2
+
+STAGE PLANS:
+ Stage: Stage-2
+ Spark
+#### A masked pattern was here ####
+ Vertices:
+ Map 3
+ Map Operator Tree:
+ TableScan
+ alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
+
+ Stage: Stage-4
+ Spark
+ Edges:
+ Reducer 4 <- Map 1 (GROUP SORT)
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ Select Operator
+ expressions: key (type: string), value (type: string)
+ outputColumnNames: key, value
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Group By Operator
+ aggregations: count(DISTINCT value)
+ keys: key (type: string), value (type: string)
+ mode: hash
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: string), _col1 (type: string)
+ sort order: ++
+ Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Reducer 4
+ Reduce Operator Tree:
+ Group By Operator
+ aggregations: count(DISTINCT KEY._col1:0._col0)
+ keys: KEY._col0 (type: string)
+ mode: mergepartial
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: string), UDFToDouble(_col1) (type: double)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e1
+
+ Stage: Stage-3
+ Dependency Collection
+
+ Stage: Stage-0
+ Move Operator
+ tables:
+ replace: true
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e1
+
+ Stage: Stage-6
+ Stats-Aggr Operator
+
+ Stage: Stage-1
+ Move Operator
+ tables:
+ replace: true
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e3
+
+ Stage: Stage-7
+ Stats-Aggr Operator
+
+ Stage: Stage-5
+ Spark
+ Edges:
+ Reducer 5 <- Map 2 (GROUP SORT)
+#### A masked pattern was here ####
+ Vertices:
+ Map 2
+ Map Operator Tree:
+ TableScan
+ Select Operator
+ expressions: value (type: string), key (type: string)
+ outputColumnNames: value, key
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Group By Operator
+ aggregations: count(DISTINCT key)
+ keys: value (type: string), key (type: string)
+ mode: hash
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: string), _col1 (type: string)
+ sort order: ++
+ Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Reducer 5
+ Reduce Operator Tree:
+ Group By Operator
+ aggregations: count(DISTINCT KEY._col1:0._col0)
+ keys: KEY._col0 (type: string)
+ mode: mergepartial
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: string), UDFToDouble(_col1) (type: double)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e3
+
+PREHOOK: query: explain
+FROM (select key, cast(key as double) as keyD, value from src order by key) a
+INSERT OVERWRITE TABLE e1
+ SELECT key, COUNT(distinct value) group by key
+INSERT OVERWRITE TABLE e2
+ SELECT key, sum(keyD), value group by key, value
+INSERT overwrite table e3
+ SELECT key, COUNT(distinct keyD) group by key, keyD, value
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+FROM (select key, cast(key as double) as keyD, value from src order by key) a
+INSERT OVERWRITE TABLE e1
+ SELECT key, COUNT(distinct value) group by key
+INSERT OVERWRITE TABLE e2
+ SELECT key, sum(keyD), value group by key, value
+INSERT overwrite table e3
+ SELECT key, COUNT(distinct keyD) group by key, keyD, value
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+ Stage-3 is a root stage
+ Stage-5 depends on stages: Stage-3
+ Stage-4 depends on stages: Stage-5, Stage-6
+ Stage-0 depends on stages: Stage-4
+ Stage-7 depends on stages: Stage-0
+ Stage-1 depends on stages: Stage-4
+ Stage-8 depends on stages: Stage-1
+ Stage-2 depends on stages: Stage-4
+ Stage-9 depends on stages: Stage-2
+ Stage-6 depends on stages: Stage-3
+
+STAGE PLANS:
+ Stage: Stage-3
+ Spark
+ Edges:
+ Reducer 4 <- Map 3 (GROUP SORT)
+#### A masked pattern was here ####
+ Vertices:
+ Map 3
+ Map Operator Tree:
+ TableScan
+ alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string), UDFToDouble(key) (type: double), value (type: string)
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: string), _col1 (type: double), _col2 (type: string)
+ sort order: ++++
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: double), _col2 (type: string)
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col1 (type: double), _col2 (type: string)
+ Reducer 4
+ Reduce Operator Tree:
+ Select Operator
+ expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: double), VALUE._col1 (type: string)
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
+
+ Stage: Stage-5
+ Spark
+ Edges:
+ Reducer 5 <- Map 1 (GROUP SORT)
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ Select Operator
+ expressions: _col0 (type: string), _col1 (type: double), _col2 (type: string)
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: string), _col2 (type: string)
+ sort order: ++
+ Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col1 (type: double)
+ Reducer 5
+ Reduce Operator Tree:
+ Forward
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Group By Operator
+ aggregations: count(DISTINCT KEY._col1:0._col0)
+ keys: KEY._col0 (type: string)
+ mode: complete
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: string), UDFToDouble(_col1) (type: double)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e1
+ Group By Operator
+ aggregations: sum(VALUE._col0)
+ keys: KEY._col0 (type: string), KEY._col1:0._col0 (type: string)
+ mode: complete
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: string), _col2 (type: double), _col1 (type: string)
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e2
+
+ Stage: Stage-4
+ Dependency Collection
+
+ Stage: Stage-0
+ Move Operator
+ tables:
+ replace: true
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e1
+
+ Stage: Stage-7
+ Stats-Aggr Operator
+
+ Stage: Stage-1
+ Move Operator
+ tables:
+ replace: true
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e2
+
+ Stage: Stage-8
+ Stats-Aggr Operator
+
+ Stage: Stage-2
+ Move Operator
+ tables:
+ replace: true
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e3
+
+ Stage: Stage-9
+ Stats-Aggr Operator
+
+ Stage: Stage-6
+ Spark
+#### A masked pattern was here ####
+ Vertices:
+ Map 2
+ Map Operator Tree:
+ TableScan
+ Select Operator
+ expressions: _col0 (type: string), _col1 (type: double), _col2 (type: string)
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Group By Operator
+ aggregations: count(DISTINCT _col1)
+ keys: _col0 (type: string), _col1 (type: double), _col2 (type: string)
+ mode: complete
+ outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: string), UDFToDouble(_col3) (type: double)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e3
+