You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by gu...@apache.org on 2017/02/03 21:51:02 UTC
[49/51] [partial] hive git commit: HIVE-15790: Remove unused beeline
golden files (Gunther Hagleitner, reviewed by Sergey Shelukhin)
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/auto_join1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/auto_join1.q.out b/ql/src/test/results/beelinepositive/auto_join1.q.out
deleted file mode 100644
index 5733fda..0000000
--- a/ql/src/test/results/beelinepositive/auto_join1.q.out
+++ /dev/null
@@ -1,246 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/auto_join1.q.raw". Enter "record" with no arguments to stop it.
->>> !run !!{qFileDirectory}!!/auto_join1.q
->>> set hive.auto.convert.join =true;
-No rows affected
->>>
->>> CREATE TABLE dest_j1(key INT, value STRING) STORED AS TEXTFILE;
-No rows affected
->>>
->>> explain
-FROM src src1 JOIN src src2 ON (src1.key = src2.key)
-INSERT OVERWRITE TABLE dest_j1 SELECT src1.key, src2.value;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_TABREF (TOK_TABNAME src) src1) (TOK_TABREF (TOK_TABNAME src) src2) (= (. (TOK_TABLE_OR_COL src1) key) (. (TOK_TABLE_OR_COL src2) key)))) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME dest_j1))) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL src1) key)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL src2) value)))))'
-''
-'STAGE DEPENDENCIES:'
-' Stage-6 is a root stage , consists of Stage-7, Stage-8, Stage-1'
-' Stage-7 has a backup stage: Stage-1'
-' Stage-4 depends on stages: Stage-7'
-' Stage-0 depends on stages: Stage-1, Stage-4, Stage-5'
-' Stage-2 depends on stages: Stage-0'
-' Stage-8 has a backup stage: Stage-1'
-' Stage-5 depends on stages: Stage-8'
-' Stage-1'
-''
-'STAGE PLANS:'
-' Stage: Stage-6'
-' Conditional Operator'
-''
-' Stage: Stage-7'
-' Map Reduce Local Work'
-' Alias -> Map Local Tables:'
-' src2 '
-' Fetch Operator'
-' limit: -1'
-' Alias -> Map Local Operator Tree:'
-' src2 '
-' TableScan'
-' alias: src2'
-' HashTable Sink Operator'
-' condition expressions:'
-' 0 {key}'
-' 1 {value}'
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[key]]'
-' 1 [Column[key]]'
-' Position of Big Table: 0'
-''
-' Stage: Stage-4'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src1 '
-' TableScan'
-' alias: src1'
-' Map Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' condition expressions:'
-' 0 {key}'
-' 1 {value}'
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[key]]'
-' 1 [Column[key]]'
-' outputColumnNames: _col0, _col5'
-' Position of Big Table: 0'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col5'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Select Operator'
-' expressions:'
-' expr: UDFToInteger(_col0)'
-' type: int'
-' expr: _col1'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 1'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-' name: auto_join1.dest_j1'
-' Local Work:'
-' Map Reduce Local Work'
-''
-' Stage: Stage-0'
-' Move Operator'
-' tables:'
-' replace: true'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-' name: auto_join1.dest_j1'
-''
-' Stage: Stage-2'
-' Stats-Aggr Operator'
-''
-' Stage: Stage-8'
-' Map Reduce Local Work'
-' Alias -> Map Local Tables:'
-' src1 '
-' Fetch Operator'
-' limit: -1'
-' Alias -> Map Local Operator Tree:'
-' src1 '
-' TableScan'
-' alias: src1'
-' HashTable Sink Operator'
-' condition expressions:'
-' 0 {key}'
-' 1 {value}'
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[key]]'
-' 1 [Column[key]]'
-' Position of Big Table: 1'
-''
-' Stage: Stage-5'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src2 '
-' TableScan'
-' alias: src2'
-' Map Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' condition expressions:'
-' 0 {key}'
-' 1 {value}'
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[key]]'
-' 1 [Column[key]]'
-' outputColumnNames: _col0, _col5'
-' Position of Big Table: 1'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col5'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Select Operator'
-' expressions:'
-' expr: UDFToInteger(_col0)'
-' type: int'
-' expr: _col1'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 1'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-' name: auto_join1.dest_j1'
-' Local Work:'
-' Map Reduce Local Work'
-''
-' Stage: Stage-1'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src1 '
-' TableScan'
-' alias: src1'
-' Reduce Output Operator'
-' key expressions:'
-' expr: key'
-' type: string'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: key'
-' type: string'
-' tag: 0'
-' value expressions:'
-' expr: key'
-' type: string'
-' src2 '
-' TableScan'
-' alias: src2'
-' Reduce Output Operator'
-' key expressions:'
-' expr: key'
-' type: string'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: key'
-' type: string'
-' tag: 1'
-' value expressions:'
-' expr: value'
-' type: string'
-' Reduce Operator Tree:'
-' Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' condition expressions:'
-' 0 {VALUE._col0}'
-' 1 {VALUE._col1}'
-' handleSkewJoin: false'
-' outputColumnNames: _col0, _col5'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col5'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Select Operator'
-' expressions:'
-' expr: UDFToInteger(_col0)'
-' type: int'
-' expr: _col1'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 1'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-' name: auto_join1.dest_j1'
-''
-''
-222 rows selected
->>>
->>> FROM src src1 JOIN src src2 ON (src1.key = src2.key)
-INSERT OVERWRITE TABLE dest_j1 SELECT src1.key, src2.value;
-'_col0','_col1'
-No rows selected
->>>
->>> SELECT sum(hash(dest_j1.key,dest_j1.value)) FROM dest_j1;
-'_c0'
-'101861029915'
-1 row selected
->>> !record
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/auto_join10.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/auto_join10.q.out b/ql/src/test/results/beelinepositive/auto_join10.q.out
deleted file mode 100644
index ffd7b4b..0000000
--- a/ql/src/test/results/beelinepositive/auto_join10.q.out
+++ /dev/null
@@ -1,294 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/auto_join10.q.raw". Enter "record" with no arguments to stop it.
->>> !run !!{qFileDirectory}!!/auto_join10.q
->>> set hive.auto.convert.join = true;
-No rows affected
->>>
->>> explain
-FROM
-(SELECT src.* FROM src) x
-JOIN
-(SELECT src.* FROM src) Y
-ON (x.key = Y.key)
-select sum(hash(Y.key,Y.value));
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME src)))))) x) (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME src)))))) Y) (= (. (TOK_TABLE_OR_COL x) key) (. (TOK_TABLE_OR_COL Y) key)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION sum (TOK_FUNCTION hash (. (TOK_TABLE_OR_COL Y) key) (. (TOK_TABLE_OR_COL Y) value)))))))'
-''
-'STAGE DEPENDENCIES:'
-' Stage-6 is a root stage , consists of Stage-7, Stage-8, Stage-1'
-' Stage-7 has a backup stage: Stage-1'
-' Stage-4 depends on stages: Stage-7'
-' Stage-2 depends on stages: Stage-1, Stage-4, Stage-5'
-' Stage-8 has a backup stage: Stage-1'
-' Stage-5 depends on stages: Stage-8'
-' Stage-1'
-' Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-' Stage: Stage-6'
-' Conditional Operator'
-''
-' Stage: Stage-7'
-' Map Reduce Local Work'
-' Alias -> Map Local Tables:'
-' y:src '
-' Fetch Operator'
-' limit: -1'
-' Alias -> Map Local Operator Tree:'
-' y:src '
-' TableScan'
-' alias: src'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' HashTable Sink Operator'
-' condition expressions:'
-' 0 '
-' 1 {_col0} {_col1}'
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[_col0]]'
-' 1 [Column[_col0]]'
-' Position of Big Table: 0'
-''
-' Stage: Stage-4'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' x:src '
-' TableScan'
-' alias: src'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' outputColumnNames: _col0'
-' Map Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' condition expressions:'
-' 0 '
-' 1 {_col0} {_col1}'
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[_col0]]'
-' 1 [Column[_col0]]'
-' outputColumnNames: _col2, _col3'
-' Position of Big Table: 0'
-' Select Operator'
-' expressions:'
-' expr: _col2'
-' type: string'
-' expr: _col3'
-' type: string'
-' outputColumnNames: _col2, _col3'
-' Group By Operator'
-' aggregations:'
-' expr: sum(hash(_col2,_col3))'
-' bucketGroup: false'
-' mode: hash'
-' outputColumnNames: _col0'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-' Local Work:'
-' Map Reduce Local Work'
-''
-' Stage: Stage-2'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' file:!!{hive.exec.scratchdir}!! '
-' Reduce Output Operator'
-' sort order: '
-' tag: -1'
-' value expressions:'
-' expr: _col0'
-' type: bigint'
-' Reduce Operator Tree:'
-' Group By Operator'
-' aggregations:'
-' expr: sum(VALUE._col0)'
-' bucketGroup: false'
-' mode: mergepartial'
-' outputColumnNames: _col0'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: bigint'
-' outputColumnNames: _col0'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-8'
-' Map Reduce Local Work'
-' Alias -> Map Local Tables:'
-' x:src '
-' Fetch Operator'
-' limit: -1'
-' Alias -> Map Local Operator Tree:'
-' x:src '
-' TableScan'
-' alias: src'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' outputColumnNames: _col0'
-' HashTable Sink Operator'
-' condition expressions:'
-' 0 '
-' 1 {_col0} {_col1}'
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[_col0]]'
-' 1 [Column[_col0]]'
-' Position of Big Table: 1'
-''
-' Stage: Stage-5'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' y:src '
-' TableScan'
-' alias: src'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Map Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' condition expressions:'
-' 0 '
-' 1 {_col0} {_col1}'
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[_col0]]'
-' 1 [Column[_col0]]'
-' outputColumnNames: _col2, _col3'
-' Position of Big Table: 1'
-' Select Operator'
-' expressions:'
-' expr: _col2'
-' type: string'
-' expr: _col3'
-' type: string'
-' outputColumnNames: _col2, _col3'
-' Group By Operator'
-' aggregations:'
-' expr: sum(hash(_col2,_col3))'
-' bucketGroup: false'
-' mode: hash'
-' outputColumnNames: _col0'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-' Local Work:'
-' Map Reduce Local Work'
-''
-' Stage: Stage-1'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' x:src '
-' TableScan'
-' alias: src'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' outputColumnNames: _col0'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: _col0'
-' type: string'
-' tag: 0'
-' y:src '
-' TableScan'
-' alias: src'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: _col0'
-' type: string'
-' tag: 1'
-' value expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: string'
-' Reduce Operator Tree:'
-' Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' condition expressions:'
-' 0 '
-' 1 {VALUE._col0} {VALUE._col1}'
-' handleSkewJoin: false'
-' outputColumnNames: _col2, _col3'
-' Select Operator'
-' expressions:'
-' expr: _col2'
-' type: string'
-' expr: _col3'
-' type: string'
-' outputColumnNames: _col2, _col3'
-' Group By Operator'
-' aggregations:'
-' expr: sum(hash(_col2,_col3))'
-' bucketGroup: false'
-' mode: hash'
-' outputColumnNames: _col0'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-''
-' Stage: Stage-0'
-' Fetch Operator'
-' limit: -1'
-''
-''
-268 rows selected
->>>
->>> FROM
-(SELECT src.* FROM src) x
-JOIN
-(SELECT src.* FROM src) Y
-ON (x.key = Y.key)
-select sum(hash(Y.key,Y.value));
-'_c0'
-'103231310608'
-1 row selected
->>>
->>> !record
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/auto_join11.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/auto_join11.q.out b/ql/src/test/results/beelinepositive/auto_join11.q.out
deleted file mode 100644
index 839d70b..0000000
--- a/ql/src/test/results/beelinepositive/auto_join11.q.out
+++ /dev/null
@@ -1,318 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/auto_join11.q.raw". Enter "record" with no arguments to stop it.
->>> !run !!{qFileDirectory}!!/auto_join11.q
->>> set hive.auto.convert.join = true;
-No rows affected
->>>
->>> explain
-SELECT sum(hash(src1.c1, src2.c4))
-FROM
-(SELECT src.key as c1, src.value as c2 from src) src1
-JOIN
-(SELECT src.key as c3, src.value as c4 from src) src2
-ON src1.c1 = src2.c3 AND src1.c1 < 100;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) key) c1) (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) value) c2)))) src1) (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) key) c3) (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) value) c4)))) src2) (AND (= (. (TOK_TABLE_OR_COL src1) c1) (. (TOK_TABLE_OR_COL src2) c3)) (< (. (TOK_TABLE_OR_COL src1) c1) 100)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION sum (TOK_FUNCTION hash (. (TOK_TABLE_OR_COL src1) c1) (. (TOK_TABLE_OR_COL src2) c4)))))))'
-''
-'STAGE DEPENDENCIES:'
-' Stage-6 is a root stage , consists of Stage-7, Stage-8, Stage-1'
-' Stage-7 has a backup stage: Stage-1'
-' Stage-4 depends on stages: Stage-7'
-' Stage-2 depends on stages: Stage-1, Stage-4, Stage-5'
-' Stage-8 has a backup stage: Stage-1'
-' Stage-5 depends on stages: Stage-8'
-' Stage-1'
-' Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-' Stage: Stage-6'
-' Conditional Operator'
-''
-' Stage: Stage-7'
-' Map Reduce Local Work'
-' Alias -> Map Local Tables:'
-' src2:src '
-' Fetch Operator'
-' limit: -1'
-' Alias -> Map Local Operator Tree:'
-' src2:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: (key < 100.0)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' HashTable Sink Operator'
-' condition expressions:'
-' 0 {_col0}'
-' 1 {_col1}'
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[_col0]]'
-' 1 [Column[_col0]]'
-' Position of Big Table: 0'
-''
-' Stage: Stage-4'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src1:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: (key < 100.0)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' outputColumnNames: _col0'
-' Map Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' condition expressions:'
-' 0 {_col0}'
-' 1 {_col1}'
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[_col0]]'
-' 1 [Column[_col0]]'
-' outputColumnNames: _col0, _col3'
-' Position of Big Table: 0'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col3'
-' type: string'
-' outputColumnNames: _col0, _col3'
-' Group By Operator'
-' aggregations:'
-' expr: sum(hash(_col0,_col3))'
-' bucketGroup: false'
-' mode: hash'
-' outputColumnNames: _col0'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-' Local Work:'
-' Map Reduce Local Work'
-''
-' Stage: Stage-2'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' file:!!{hive.exec.scratchdir}!! '
-' Reduce Output Operator'
-' sort order: '
-' tag: -1'
-' value expressions:'
-' expr: _col0'
-' type: bigint'
-' Reduce Operator Tree:'
-' Group By Operator'
-' aggregations:'
-' expr: sum(VALUE._col0)'
-' bucketGroup: false'
-' mode: mergepartial'
-' outputColumnNames: _col0'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: bigint'
-' outputColumnNames: _col0'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-8'
-' Map Reduce Local Work'
-' Alias -> Map Local Tables:'
-' src1:src '
-' Fetch Operator'
-' limit: -1'
-' Alias -> Map Local Operator Tree:'
-' src1:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: (key < 100.0)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' outputColumnNames: _col0'
-' HashTable Sink Operator'
-' condition expressions:'
-' 0 {_col0}'
-' 1 {_col1}'
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[_col0]]'
-' 1 [Column[_col0]]'
-' Position of Big Table: 1'
-''
-' Stage: Stage-5'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src2:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: (key < 100.0)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Map Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' condition expressions:'
-' 0 {_col0}'
-' 1 {_col1}'
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[_col0]]'
-' 1 [Column[_col0]]'
-' outputColumnNames: _col0, _col3'
-' Position of Big Table: 1'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col3'
-' type: string'
-' outputColumnNames: _col0, _col3'
-' Group By Operator'
-' aggregations:'
-' expr: sum(hash(_col0,_col3))'
-' bucketGroup: false'
-' mode: hash'
-' outputColumnNames: _col0'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-' Local Work:'
-' Map Reduce Local Work'
-''
-' Stage: Stage-1'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src1:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: (key < 100.0)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' outputColumnNames: _col0'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: _col0'
-' type: string'
-' tag: 0'
-' value expressions:'
-' expr: _col0'
-' type: string'
-' src2:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: (key < 100.0)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: _col0'
-' type: string'
-' tag: 1'
-' value expressions:'
-' expr: _col1'
-' type: string'
-' Reduce Operator Tree:'
-' Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' condition expressions:'
-' 0 {VALUE._col0}'
-' 1 {VALUE._col1}'
-' handleSkewJoin: false'
-' outputColumnNames: _col0, _col3'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col3'
-' type: string'
-' outputColumnNames: _col0, _col3'
-' Group By Operator'
-' aggregations:'
-' expr: sum(hash(_col0,_col3))'
-' bucketGroup: false'
-' mode: hash'
-' outputColumnNames: _col0'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-''
-' Stage: Stage-0'
-' Fetch Operator'
-' limit: -1'
-''
-''
-293 rows selected
->>>
->>> SELECT sum(hash(src1.c1, src2.c4))
-FROM
-(SELECT src.key as c1, src.value as c2 from src) src1
-JOIN
-(SELECT src.key as c3, src.value as c4 from src) src2
-ON src1.c1 = src2.c3 AND src1.c1 < 100;
-'_c0'
-'-101333194320'
-1 row selected
->>> !record
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/auto_join12.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/auto_join12.q.out b/ql/src/test/results/beelinepositive/auto_join12.q.out
deleted file mode 100644
index 7083a19..0000000
--- a/ql/src/test/results/beelinepositive/auto_join12.q.out
+++ /dev/null
@@ -1,525 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/auto_join12.q.raw". Enter "record" with no arguments to stop it.
->>> !run !!{qFileDirectory}!!/auto_join12.q
->>>
->>>
->>> set hive.auto.convert.join = true;
-No rows affected
->>>
->>>
->>> explain
-SELECT sum(hash(src1.c1, src2.c4))
-FROM
-(SELECT src.key as c1, src.value as c2 from src) src1
-JOIN
-(SELECT src.key as c3, src.value as c4 from src) src2
-ON src1.c1 = src2.c3 AND src1.c1 < 100
-JOIN
-(SELECT src.key as c5, src.value as c6 from src) src3
-ON src1.c1 = src3.c5 AND src3.c5 < 80;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_JOIN (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) key) c1) (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) value) c2)))) src1) (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) key) c3) (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) value) c4)))) src2) (AND (= (. (TOK_TABLE_OR_COL src1) c1) (. (TOK_TABLE_OR_COL src2) c3)) (< (. (TOK_TABLE_OR_COL src1) c1) 100))) (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) key) c5) (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) value) c6)))) src3) (AND (= (. (TOK_TABLE_OR_COL src1) c1) (. (TOK_TABLE_OR_COL src3) c5)) (< (. (TOK_TABLE_OR_COL src3) c5) 80)))) (TOK_INSERT
(TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION sum (TOK_FUNCTION hash (. (TOK_TABLE_OR_COL src1) c1) (. (TOK_TABLE_OR_COL src2) c4)))))))'
-''
-'STAGE DEPENDENCIES:'
-' Stage-8 is a root stage , consists of Stage-9, Stage-10, Stage-11, Stage-1'
-' Stage-9 has a backup stage: Stage-1'
-' Stage-5 depends on stages: Stage-9'
-' Stage-2 depends on stages: Stage-1, Stage-5, Stage-6, Stage-7'
-' Stage-10 has a backup stage: Stage-1'
-' Stage-6 depends on stages: Stage-10'
-' Stage-11 has a backup stage: Stage-1'
-' Stage-7 depends on stages: Stage-11'
-' Stage-1'
-' Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-' Stage: Stage-8'
-' Conditional Operator'
-''
-' Stage: Stage-9'
-' Map Reduce Local Work'
-' Alias -> Map Local Tables:'
-' src2:src '
-' Fetch Operator'
-' limit: -1'
-' src3:src '
-' Fetch Operator'
-' limit: -1'
-' Alias -> Map Local Operator Tree:'
-' src2:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: ((key < 100) and (key < 80))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' HashTable Sink Operator'
-' condition expressions:'
-' 0 {_col0}'
-' 1 {_col1}'
-' 2 '
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[_col0]]'
-' 1 [Column[_col0]]'
-' 2 [Column[_col0]]'
-' Position of Big Table: 0'
-' src3:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: ((key < 80) and (key < 100))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' outputColumnNames: _col0'
-' HashTable Sink Operator'
-' condition expressions:'
-' 0 {_col0}'
-' 1 {_col1}'
-' 2 '
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[_col0]]'
-' 1 [Column[_col0]]'
-' 2 [Column[_col0]]'
-' Position of Big Table: 0'
-''
-' Stage: Stage-5'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src1:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: ((key < 100) and (key < 80))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' outputColumnNames: _col0'
-' Map Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' Inner Join 0 to 2'
-' condition expressions:'
-' 0 {_col0}'
-' 1 {_col1}'
-' 2 '
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[_col0]]'
-' 1 [Column[_col0]]'
-' 2 [Column[_col0]]'
-' outputColumnNames: _col0, _col3'
-' Position of Big Table: 0'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col3'
-' type: string'
-' outputColumnNames: _col0, _col3'
-' Group By Operator'
-' aggregations:'
-' expr: sum(hash(_col0,_col3))'
-' bucketGroup: false'
-' mode: hash'
-' outputColumnNames: _col0'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-' Local Work:'
-' Map Reduce Local Work'
-''
-' Stage: Stage-2'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' file:!!{hive.exec.scratchdir}!! '
-' Reduce Output Operator'
-' sort order: '
-' tag: -1'
-' value expressions:'
-' expr: _col0'
-' type: bigint'
-' Reduce Operator Tree:'
-' Group By Operator'
-' aggregations:'
-' expr: sum(VALUE._col0)'
-' bucketGroup: false'
-' mode: mergepartial'
-' outputColumnNames: _col0'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: bigint'
-' outputColumnNames: _col0'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-10'
-' Map Reduce Local Work'
-' Alias -> Map Local Tables:'
-' src1:src '
-' Fetch Operator'
-' limit: -1'
-' src3:src '
-' Fetch Operator'
-' limit: -1'
-' Alias -> Map Local Operator Tree:'
-' src1:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: ((key < 100) and (key < 80))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' outputColumnNames: _col0'
-' HashTable Sink Operator'
-' condition expressions:'
-' 0 {_col0}'
-' 1 {_col1}'
-' 2 '
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[_col0]]'
-' 1 [Column[_col0]]'
-' 2 [Column[_col0]]'
-' Position of Big Table: 1'
-' src3:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: ((key < 80) and (key < 100))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' outputColumnNames: _col0'
-' HashTable Sink Operator'
-' condition expressions:'
-' 0 {_col0}'
-' 1 {_col1}'
-' 2 '
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[_col0]]'
-' 1 [Column[_col0]]'
-' 2 [Column[_col0]]'
-' Position of Big Table: 1'
-''
-' Stage: Stage-6'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src2:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: ((key < 100) and (key < 80))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Map Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' Inner Join 0 to 2'
-' condition expressions:'
-' 0 {_col0}'
-' 1 {_col1}'
-' 2 '
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[_col0]]'
-' 1 [Column[_col0]]'
-' 2 [Column[_col0]]'
-' outputColumnNames: _col0, _col3'
-' Position of Big Table: 1'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col3'
-' type: string'
-' outputColumnNames: _col0, _col3'
-' Group By Operator'
-' aggregations:'
-' expr: sum(hash(_col0,_col3))'
-' bucketGroup: false'
-' mode: hash'
-' outputColumnNames: _col0'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-' Local Work:'
-' Map Reduce Local Work'
-''
-' Stage: Stage-11'
-' Map Reduce Local Work'
-' Alias -> Map Local Tables:'
-' src1:src '
-' Fetch Operator'
-' limit: -1'
-' src2:src '
-' Fetch Operator'
-' limit: -1'
-' Alias -> Map Local Operator Tree:'
-' src1:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: ((key < 100) and (key < 80))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' outputColumnNames: _col0'
-' HashTable Sink Operator'
-' condition expressions:'
-' 0 {_col0}'
-' 1 {_col1}'
-' 2 '
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[_col0]]'
-' 1 [Column[_col0]]'
-' 2 [Column[_col0]]'
-' Position of Big Table: 2'
-' src2:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: ((key < 100) and (key < 80))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' HashTable Sink Operator'
-' condition expressions:'
-' 0 {_col0}'
-' 1 {_col1}'
-' 2 '
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[_col0]]'
-' 1 [Column[_col0]]'
-' 2 [Column[_col0]]'
-' Position of Big Table: 2'
-''
-' Stage: Stage-7'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src3:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: ((key < 80) and (key < 100))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' outputColumnNames: _col0'
-' Map Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' Inner Join 0 to 2'
-' condition expressions:'
-' 0 {_col0}'
-' 1 {_col1}'
-' 2 '
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[_col0]]'
-' 1 [Column[_col0]]'
-' 2 [Column[_col0]]'
-' outputColumnNames: _col0, _col3'
-' Position of Big Table: 2'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col3'
-' type: string'
-' outputColumnNames: _col0, _col3'
-' Group By Operator'
-' aggregations:'
-' expr: sum(hash(_col0,_col3))'
-' bucketGroup: false'
-' mode: hash'
-' outputColumnNames: _col0'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-' Local Work:'
-' Map Reduce Local Work'
-''
-' Stage: Stage-1'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src1:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: ((key < 100) and (key < 80))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' outputColumnNames: _col0'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: _col0'
-' type: string'
-' tag: 0'
-' value expressions:'
-' expr: _col0'
-' type: string'
-' src2:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: ((key < 100) and (key < 80))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: _col0'
-' type: string'
-' tag: 1'
-' value expressions:'
-' expr: _col1'
-' type: string'
-' src3:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: ((key < 80) and (key < 100))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' outputColumnNames: _col0'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: _col0'
-' type: string'
-' tag: 2'
-' Reduce Operator Tree:'
-' Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' Inner Join 0 to 2'
-' condition expressions:'
-' 0 {VALUE._col0}'
-' 1 {VALUE._col1}'
-' 2 '
-' handleSkewJoin: false'
-' outputColumnNames: _col0, _col3'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col3'
-' type: string'
-' outputColumnNames: _col0, _col3'
-' Group By Operator'
-' aggregations:'
-' expr: sum(hash(_col0,_col3))'
-' bucketGroup: false'
-' mode: hash'
-' outputColumnNames: _col0'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-''
-' Stage: Stage-0'
-' Fetch Operator'
-' limit: -1'
-''
-''
-491 rows selected
->>>
->>> SELECT sum(hash(src1.c1, src2.c4))
-FROM
-(SELECT src.key as c1, src.value as c2 from src) src1
-JOIN
-(SELECT src.key as c3, src.value as c4 from src) src2
-ON src1.c1 = src2.c3 AND src1.c1 < 100
-JOIN
-(SELECT src.key as c5, src.value as c6 from src) src3
-ON src1.c1 = src3.c5 AND src3.c5 < 80;
-'_c0'
-'-136843922952'
-1 row selected
->>> !record
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/auto_join13.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/auto_join13.q.out b/ql/src/test/results/beelinepositive/auto_join13.q.out
deleted file mode 100644
index 627b2c3..0000000
--- a/ql/src/test/results/beelinepositive/auto_join13.q.out
+++ /dev/null
@@ -1,499 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/auto_join13.q.raw". Enter "record" with no arguments to stop it.
->>> !run !!{qFileDirectory}!!/auto_join13.q
->>>
->>> set hive.auto.convert.join = true;
-No rows affected
->>>
->>> explain
-SELECT sum(hash(src1.c1, src2.c4))
-FROM
-(SELECT src.key as c1, src.value as c2 from src) src1
-JOIN
-(SELECT src.key as c3, src.value as c4 from src) src2
-ON src1.c1 = src2.c3 AND src1.c1 < 100
-JOIN
-(SELECT src.key as c5, src.value as c6 from src) src3
-ON src1.c1 + src2.c3 = src3.c5 AND src3.c5 < 200;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_JOIN (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) key) c1) (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) value) c2)))) src1) (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) key) c3) (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) value) c4)))) src2) (AND (= (. (TOK_TABLE_OR_COL src1) c1) (. (TOK_TABLE_OR_COL src2) c3)) (< (. (TOK_TABLE_OR_COL src1) c1) 100))) (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) key) c5) (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) value) c6)))) src3) (AND (= (+ (. (TOK_TABLE_OR_COL src1) c1) (. (TOK_TABLE_OR_COL src2) c3)) (. (TOK_TABLE_OR_COL src3) c5)) (< (. (TOK_TABLE_O
R_COL src3) c5) 200)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION sum (TOK_FUNCTION hash (. (TOK_TABLE_OR_COL src1) c1) (. (TOK_TABLE_OR_COL src2) c4)))))))'
-''
-'STAGE DEPENDENCIES:'
-' Stage-11 is a root stage , consists of Stage-14, Stage-15, Stage-1'
-' Stage-14 has a backup stage: Stage-1'
-' Stage-9 depends on stages: Stage-14'
-' Stage-8 depends on stages: Stage-1, Stage-9, Stage-10 , consists of Stage-12, Stage-13, Stage-2'
-' Stage-12 has a backup stage: Stage-2'
-' Stage-6 depends on stages: Stage-12'
-' Stage-3 depends on stages: Stage-2, Stage-6, Stage-7'
-' Stage-13 has a backup stage: Stage-2'
-' Stage-7 depends on stages: Stage-13'
-' Stage-2'
-' Stage-15 has a backup stage: Stage-1'
-' Stage-10 depends on stages: Stage-15'
-' Stage-1'
-' Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-' Stage: Stage-11'
-' Conditional Operator'
-''
-' Stage: Stage-14'
-' Map Reduce Local Work'
-' Alias -> Map Local Tables:'
-' src2:src '
-' Fetch Operator'
-' limit: -1'
-' Alias -> Map Local Operator Tree:'
-' src2:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: (key < 100)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' HashTable Sink Operator'
-' condition expressions:'
-' 0 {_col0}'
-' 1 {_col0} {_col1}'
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[_col0]]'
-' 1 [Column[_col0]]'
-' Position of Big Table: 0'
-''
-' Stage: Stage-9'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src1:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: (key < 100)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' outputColumnNames: _col0'
-' Map Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' condition expressions:'
-' 0 {_col0}'
-' 1 {_col0} {_col1}'
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[_col0]]'
-' 1 [Column[_col0]]'
-' outputColumnNames: _col0, _col2, _col3'
-' Position of Big Table: 0'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-' Local Work:'
-' Map Reduce Local Work'
-''
-' Stage: Stage-8'
-' Conditional Operator'
-''
-' Stage: Stage-12'
-' Map Reduce Local Work'
-' Alias -> Map Local Tables:'
-' src3:src '
-' Fetch Operator'
-' limit: -1'
-' Alias -> Map Local Operator Tree:'
-' src3:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: (key < 200)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' outputColumnNames: _col0'
-' HashTable Sink Operator'
-' condition expressions:'
-' 0 {_col3} {_col0}'
-' 1 '
-' handleSkewJoin: false'
-' keys:'
-' 0 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0], Column[_col2]()]'
-' 1 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0]()]'
-' Position of Big Table: 0'
-''
-' Stage: Stage-6'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' $INTNAME '
-' Map Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' condition expressions:'
-' 0 {_col3} {_col0}'
-' 1 '
-' handleSkewJoin: false'
-' keys:'
-' 0 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0], Column[_col2]()]'
-' 1 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0]()]'
-' outputColumnNames: _col1, _col2'
-' Position of Big Table: 0'
-' Select Operator'
-' expressions:'
-' expr: _col2'
-' type: string'
-' expr: _col1'
-' type: string'
-' outputColumnNames: _col2, _col1'
-' Group By Operator'
-' aggregations:'
-' expr: sum(hash(_col2,_col1))'
-' bucketGroup: false'
-' mode: hash'
-' outputColumnNames: _col0'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-' Local Work:'
-' Map Reduce Local Work'
-''
-' Stage: Stage-3'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' file:!!{hive.exec.scratchdir}!! '
-' Reduce Output Operator'
-' sort order: '
-' tag: -1'
-' value expressions:'
-' expr: _col0'
-' type: bigint'
-' Reduce Operator Tree:'
-' Group By Operator'
-' aggregations:'
-' expr: sum(VALUE._col0)'
-' bucketGroup: false'
-' mode: mergepartial'
-' outputColumnNames: _col0'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: bigint'
-' outputColumnNames: _col0'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-13'
-' Map Reduce Local Work'
-' Alias -> Map Local Tables:'
-' $INTNAME '
-' Fetch Operator'
-' limit: -1'
-' Alias -> Map Local Operator Tree:'
-' $INTNAME '
-' HashTable Sink Operator'
-' condition expressions:'
-' 0 {_col3} {_col0}'
-' 1 '
-' handleSkewJoin: false'
-' keys:'
-' 0 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0], Column[_col2]()]'
-' 1 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0]()]'
-' Position of Big Table: 1'
-''
-' Stage: Stage-7'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src3:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: (key < 200)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' outputColumnNames: _col0'
-' Map Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' condition expressions:'
-' 0 {_col3} {_col0}'
-' 1 '
-' handleSkewJoin: false'
-' keys:'
-' 0 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0], Column[_col2]()]'
-' 1 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0]()]'
-' outputColumnNames: _col1, _col2'
-' Position of Big Table: 1'
-' Select Operator'
-' expressions:'
-' expr: _col2'
-' type: string'
-' expr: _col1'
-' type: string'
-' outputColumnNames: _col2, _col1'
-' Group By Operator'
-' aggregations:'
-' expr: sum(hash(_col2,_col1))'
-' bucketGroup: false'
-' mode: hash'
-' outputColumnNames: _col0'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-' Local Work:'
-' Map Reduce Local Work'
-''
-' Stage: Stage-2'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' $INTNAME '
-' Reduce Output Operator'
-' key expressions:'
-' expr: (_col0 + _col2)'
-' type: double'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: (_col0 + _col2)'
-' type: double'
-' tag: 0'
-' value expressions:'
-' expr: _col3'
-' type: string'
-' expr: _col0'
-' type: string'
-' src3:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: (key < 200)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' outputColumnNames: _col0'
-' Reduce Output Operator'
-' key expressions:'
-' expr: UDFToDouble(_col0)'
-' type: double'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: UDFToDouble(_col0)'
-' type: double'
-' tag: 1'
-' Reduce Operator Tree:'
-' Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' condition expressions:'
-' 0 {VALUE._col1} {VALUE._col2}'
-' 1 '
-' handleSkewJoin: false'
-' outputColumnNames: _col1, _col2'
-' Select Operator'
-' expressions:'
-' expr: _col2'
-' type: string'
-' expr: _col1'
-' type: string'
-' outputColumnNames: _col2, _col1'
-' Group By Operator'
-' aggregations:'
-' expr: sum(hash(_col2,_col1))'
-' bucketGroup: false'
-' mode: hash'
-' outputColumnNames: _col0'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-''
-' Stage: Stage-15'
-' Map Reduce Local Work'
-' Alias -> Map Local Tables:'
-' src1:src '
-' Fetch Operator'
-' limit: -1'
-' Alias -> Map Local Operator Tree:'
-' src1:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: (key < 100)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' outputColumnNames: _col0'
-' HashTable Sink Operator'
-' condition expressions:'
-' 0 {_col0}'
-' 1 {_col0} {_col1}'
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[_col0]]'
-' 1 [Column[_col0]]'
-' Position of Big Table: 1'
-''
-' Stage: Stage-10'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src2:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: (key < 100)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Map Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' condition expressions:'
-' 0 {_col0}'
-' 1 {_col0} {_col1}'
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[_col0]]'
-' 1 [Column[_col0]]'
-' outputColumnNames: _col0, _col2, _col3'
-' Position of Big Table: 1'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-' Local Work:'
-' Map Reduce Local Work'
-''
-' Stage: Stage-1'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src1:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: (key < 100)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' outputColumnNames: _col0'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: _col0'
-' type: string'
-' tag: 0'
-' value expressions:'
-' expr: _col0'
-' type: string'
-' src2:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: (key < 100)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: _col0'
-' type: string'
-' tag: 1'
-' value expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: string'
-' Reduce Operator Tree:'
-' Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' condition expressions:'
-' 0 {VALUE._col0}'
-' 1 {VALUE._col0} {VALUE._col1}'
-' handleSkewJoin: false'
-' outputColumnNames: _col0, _col2, _col3'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-''
-' Stage: Stage-0'
-' Fetch Operator'
-' limit: -1'
-''
-''
-467 rows selected
->>>
->>> SELECT sum(hash(src1.c1, src2.c4))
-FROM
-(SELECT src.key as c1, src.value as c2 from src) src1
-JOIN
-(SELECT src.key as c3, src.value as c4 from src) src2
-ON src1.c1 = src2.c3 AND src1.c1 < 100
-JOIN
-(SELECT src.key as c5, src.value as c6 from src) src3
-ON src1.c1 + src2.c3 = src3.c5 AND src3.c5 < 200;
-'_c0'
-'-97670109576'
-1 row selected
->>> !record
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/auto_join14.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/auto_join14.q.out b/ql/src/test/results/beelinepositive/auto_join14.q.out
deleted file mode 100644
index 46d9ca5..0000000
--- a/ql/src/test/results/beelinepositive/auto_join14.q.out
+++ /dev/null
@@ -1,276 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/auto_join14.q.raw". Enter "record" with no arguments to stop it.
->>> !run !!{qFileDirectory}!!/auto_join14.q
->>>
->>> set hive.auto.convert.join = true;
-No rows affected
->>>
->>> CREATE TABLE dest1(c1 INT, c2 STRING) STORED AS TEXTFILE;
-No rows affected
->>>
->>> set mapred.job.tracker=does.notexist.com:666;
-No rows affected
->>> set hive.exec.mode.local.auto=true;
-No rows affected
->>>
->>> explain
-FROM src JOIN srcpart ON src.key = srcpart.key AND srcpart.ds = '2008-04-08' and src.key > 100
-INSERT OVERWRITE TABLE dest1 SELECT src.key, srcpart.value;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_TABREF (TOK_TABNAME src)) (TOK_TABREF (TOK_TABNAME srcpart)) (and (AND (= (. (TOK_TABLE_OR_COL src) key) (. (TOK_TABLE_OR_COL srcpart) key)) (= (. (TOK_TABLE_OR_COL srcpart) ds) '2008-04-08')) (> (. (TOK_TABLE_OR_COL src) key) 100)))) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME dest1))) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) key)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL srcpart) value)))))'
-''
-'STAGE DEPENDENCIES:'
-' Stage-6 is a root stage , consists of Stage-7, Stage-8, Stage-1'
-' Stage-7 has a backup stage: Stage-1'
-' Stage-4 depends on stages: Stage-7'
-' Stage-0 depends on stages: Stage-1, Stage-4, Stage-5'
-' Stage-2 depends on stages: Stage-0'
-' Stage-8 has a backup stage: Stage-1'
-' Stage-5 depends on stages: Stage-8'
-' Stage-1'
-''
-'STAGE PLANS:'
-' Stage: Stage-6'
-' Conditional Operator'
-''
-' Stage: Stage-7'
-' Map Reduce Local Work'
-' Alias -> Map Local Tables:'
-' srcpart '
-' Fetch Operator'
-' limit: -1'
-' Alias -> Map Local Operator Tree:'
-' srcpart '
-' TableScan'
-' alias: srcpart'
-' Filter Operator'
-' predicate:'
-' expr: (key > 100.0)'
-' type: boolean'
-' HashTable Sink Operator'
-' condition expressions:'
-' 0 {key}'
-' 1 {value}'
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[key]]'
-' 1 [Column[key]]'
-' Position of Big Table: 0'
-''
-' Stage: Stage-4'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: (key > 100.0)'
-' type: boolean'
-' Map Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' condition expressions:'
-' 0 {key}'
-' 1 {value}'
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[key]]'
-' 1 [Column[key]]'
-' outputColumnNames: _col0, _col5'
-' Position of Big Table: 0'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col5'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Select Operator'
-' expressions:'
-' expr: UDFToInteger(_col0)'
-' type: int'
-' expr: _col1'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 1'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-' name: auto_join14.dest1'
-' Local Work:'
-' Map Reduce Local Work'
-''
-' Stage: Stage-0'
-' Move Operator'
-' tables:'
-' replace: true'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-' name: auto_join14.dest1'
-''
-' Stage: Stage-2'
-' Stats-Aggr Operator'
-''
-' Stage: Stage-8'
-' Map Reduce Local Work'
-' Alias -> Map Local Tables:'
-' src '
-' Fetch Operator'
-' limit: -1'
-' Alias -> Map Local Operator Tree:'
-' src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: (key > 100.0)'
-' type: boolean'
-' HashTable Sink Operator'
-' condition expressions:'
-' 0 {key}'
-' 1 {value}'
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[key]]'
-' 1 [Column[key]]'
-' Position of Big Table: 1'
-''
-' Stage: Stage-5'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' srcpart '
-' TableScan'
-' alias: srcpart'
-' Filter Operator'
-' predicate:'
-' expr: (key > 100.0)'
-' type: boolean'
-' Map Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' condition expressions:'
-' 0 {key}'
-' 1 {value}'
-' handleSkewJoin: false'
-' keys:'
-' 0 [Column[key]]'
-' 1 [Column[key]]'
-' outputColumnNames: _col0, _col5'
-' Position of Big Table: 1'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col5'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Select Operator'
-' expressions:'
-' expr: UDFToInteger(_col0)'
-' type: int'
-' expr: _col1'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 1'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-' name: auto_join14.dest1'
-' Local Work:'
-' Map Reduce Local Work'
-''
-' Stage: Stage-1'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: (key > 100.0)'
-' type: boolean'
-' Reduce Output Operator'
-' key expressions:'
-' expr: key'
-' type: string'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: key'
-' type: string'
-' tag: 0'
-' value expressions:'
-' expr: key'
-' type: string'
-' srcpart '
-' TableScan'
-' alias: srcpart'
-' Filter Operator'
-' predicate:'
-' expr: (key > 100.0)'
-' type: boolean'
-' Reduce Output Operator'
-' key expressions:'
-' expr: key'
-' type: string'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: key'
-' type: string'
-' tag: 1'
-' value expressions:'
-' expr: value'
-' type: string'
-' Reduce Operator Tree:'
-' Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' condition expressions:'
-' 0 {VALUE._col0}'
-' 1 {VALUE._col1}'
-' handleSkewJoin: false'
-' outputColumnNames: _col0, _col5'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col5'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Select Operator'
-' expressions:'
-' expr: UDFToInteger(_col0)'
-' type: int'
-' expr: _col1'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 1'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-' name: auto_join14.dest1'
-''
-''
-246 rows selected
->>>
->>> FROM src JOIN srcpart ON src.key = srcpart.key AND srcpart.ds = '2008-04-08' and src.key > 100
-INSERT OVERWRITE TABLE dest1 SELECT src.key, srcpart.value;
-'_col0','_col1'
-No rows selected
->>>
->>> SELECT sum(hash(dest1.c1,dest1.c2)) FROM dest1;
-'_c0'
-'404554174174'
-1 row selected
->>> !record