You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by gu...@apache.org on 2017/02/03 21:51:02 UTC

[49/51] [partial] hive git commit: HIVE-15790: Remove unused beeline golden files (Gunther Hagleitner, reviewed by Sergey Shelukhin)

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/auto_join1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/auto_join1.q.out b/ql/src/test/results/beelinepositive/auto_join1.q.out
deleted file mode 100644
index 5733fda..0000000
--- a/ql/src/test/results/beelinepositive/auto_join1.q.out
+++ /dev/null
@@ -1,246 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/auto_join1.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/auto_join1.q
->>>  set hive.auto.convert.join =true;
-No rows affected 
->>>  
->>>  CREATE TABLE dest_j1(key INT, value STRING) STORED AS TEXTFILE;
-No rows affected 
->>>  
->>>  explain 
-FROM src src1 JOIN src src2 ON (src1.key = src2.key) 
-INSERT OVERWRITE TABLE dest_j1 SELECT src1.key, src2.value;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_TABREF (TOK_TABNAME src) src1) (TOK_TABREF (TOK_TABNAME src) src2) (= (. (TOK_TABLE_OR_COL src1) key) (. (TOK_TABLE_OR_COL src2) key)))) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME dest_j1))) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL src1) key)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL src2) value)))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-6 is a root stage , consists of Stage-7, Stage-8, Stage-1'
-'  Stage-7 has a backup stage: Stage-1'
-'  Stage-4 depends on stages: Stage-7'
-'  Stage-0 depends on stages: Stage-1, Stage-4, Stage-5'
-'  Stage-2 depends on stages: Stage-0'
-'  Stage-8 has a backup stage: Stage-1'
-'  Stage-5 depends on stages: Stage-8'
-'  Stage-1'
-''
-'STAGE PLANS:'
-'  Stage: Stage-6'
-'    Conditional Operator'
-''
-'  Stage: Stage-7'
-'    Map Reduce Local Work'
-'      Alias -> Map Local Tables:'
-'        src2 '
-'          Fetch Operator'
-'            limit: -1'
-'      Alias -> Map Local Operator Tree:'
-'        src2 '
-'          TableScan'
-'            alias: src2'
-'            HashTable Sink Operator'
-'              condition expressions:'
-'                0 {key}'
-'                1 {value}'
-'              handleSkewJoin: false'
-'              keys:'
-'                0 [Column[key]]'
-'                1 [Column[key]]'
-'              Position of Big Table: 0'
-''
-'  Stage: Stage-4'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src1 '
-'          TableScan'
-'            alias: src1'
-'            Map Join Operator'
-'              condition map:'
-'                   Inner Join 0 to 1'
-'              condition expressions:'
-'                0 {key}'
-'                1 {value}'
-'              handleSkewJoin: false'
-'              keys:'
-'                0 [Column[key]]'
-'                1 [Column[key]]'
-'              outputColumnNames: _col0, _col5'
-'              Position of Big Table: 0'
-'              Select Operator'
-'                expressions:'
-'                      expr: _col0'
-'                      type: string'
-'                      expr: _col5'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                Select Operator'
-'                  expressions:'
-'                        expr: UDFToInteger(_col0)'
-'                        type: int'
-'                        expr: _col1'
-'                        type: string'
-'                  outputColumnNames: _col0, _col1'
-'                  File Output Operator'
-'                    compressed: false'
-'                    GlobalTableId: 1'
-'                    table:'
-'                        input format: org.apache.hadoop.mapred.TextInputFormat'
-'                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                        name: auto_join1.dest_j1'
-'      Local Work:'
-'        Map Reduce Local Work'
-''
-'  Stage: Stage-0'
-'    Move Operator'
-'      tables:'
-'          replace: true'
-'          table:'
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: auto_join1.dest_j1'
-''
-'  Stage: Stage-2'
-'    Stats-Aggr Operator'
-''
-'  Stage: Stage-8'
-'    Map Reduce Local Work'
-'      Alias -> Map Local Tables:'
-'        src1 '
-'          Fetch Operator'
-'            limit: -1'
-'      Alias -> Map Local Operator Tree:'
-'        src1 '
-'          TableScan'
-'            alias: src1'
-'            HashTable Sink Operator'
-'              condition expressions:'
-'                0 {key}'
-'                1 {value}'
-'              handleSkewJoin: false'
-'              keys:'
-'                0 [Column[key]]'
-'                1 [Column[key]]'
-'              Position of Big Table: 1'
-''
-'  Stage: Stage-5'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src2 '
-'          TableScan'
-'            alias: src2'
-'            Map Join Operator'
-'              condition map:'
-'                   Inner Join 0 to 1'
-'              condition expressions:'
-'                0 {key}'
-'                1 {value}'
-'              handleSkewJoin: false'
-'              keys:'
-'                0 [Column[key]]'
-'                1 [Column[key]]'
-'              outputColumnNames: _col0, _col5'
-'              Position of Big Table: 1'
-'              Select Operator'
-'                expressions:'
-'                      expr: _col0'
-'                      type: string'
-'                      expr: _col5'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                Select Operator'
-'                  expressions:'
-'                        expr: UDFToInteger(_col0)'
-'                        type: int'
-'                        expr: _col1'
-'                        type: string'
-'                  outputColumnNames: _col0, _col1'
-'                  File Output Operator'
-'                    compressed: false'
-'                    GlobalTableId: 1'
-'                    table:'
-'                        input format: org.apache.hadoop.mapred.TextInputFormat'
-'                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                        name: auto_join1.dest_j1'
-'      Local Work:'
-'        Map Reduce Local Work'
-''
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src1 '
-'          TableScan'
-'            alias: src1'
-'            Reduce Output Operator'
-'              key expressions:'
-'                    expr: key'
-'                    type: string'
-'              sort order: +'
-'              Map-reduce partition columns:'
-'                    expr: key'
-'                    type: string'
-'              tag: 0'
-'              value expressions:'
-'                    expr: key'
-'                    type: string'
-'        src2 '
-'          TableScan'
-'            alias: src2'
-'            Reduce Output Operator'
-'              key expressions:'
-'                    expr: key'
-'                    type: string'
-'              sort order: +'
-'              Map-reduce partition columns:'
-'                    expr: key'
-'                    type: string'
-'              tag: 1'
-'              value expressions:'
-'                    expr: value'
-'                    type: string'
-'      Reduce Operator Tree:'
-'        Join Operator'
-'          condition map:'
-'               Inner Join 0 to 1'
-'          condition expressions:'
-'            0 {VALUE._col0}'
-'            1 {VALUE._col1}'
-'          handleSkewJoin: false'
-'          outputColumnNames: _col0, _col5'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col0'
-'                  type: string'
-'                  expr: _col5'
-'                  type: string'
-'            outputColumnNames: _col0, _col1'
-'            Select Operator'
-'              expressions:'
-'                    expr: UDFToInteger(_col0)'
-'                    type: int'
-'                    expr: _col1'
-'                    type: string'
-'              outputColumnNames: _col0, _col1'
-'              File Output Operator'
-'                compressed: false'
-'                GlobalTableId: 1'
-'                table:'
-'                    input format: org.apache.hadoop.mapred.TextInputFormat'
-'                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                    name: auto_join1.dest_j1'
-''
-''
-222 rows selected 
->>>  
->>>  FROM src src1 JOIN src src2 ON (src1.key = src2.key) 
-INSERT OVERWRITE TABLE dest_j1 SELECT src1.key, src2.value;
-'_col0','_col1'
-No rows selected 
->>>  
->>>  SELECT sum(hash(dest_j1.key,dest_j1.value)) FROM dest_j1;
-'_c0'
-'101861029915'
-1 row selected 
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/auto_join10.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/auto_join10.q.out b/ql/src/test/results/beelinepositive/auto_join10.q.out
deleted file mode 100644
index ffd7b4b..0000000
--- a/ql/src/test/results/beelinepositive/auto_join10.q.out
+++ /dev/null
@@ -1,294 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/auto_join10.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/auto_join10.q
->>>  set hive.auto.convert.join = true;
-No rows affected 
->>>  
->>>  explain 
-FROM 
-(SELECT src.* FROM src) x 
-JOIN 
-(SELECT src.* FROM src) Y 
-ON (x.key = Y.key) 
-select sum(hash(Y.key,Y.value));
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME src)))))) x) (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME src)))))) Y) (= (. (TOK_TABLE_OR_COL x) key) (. (TOK_TABLE_OR_COL Y) key)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION sum (TOK_FUNCTION hash (. (TOK_TABLE_OR_COL Y) key) (. (TOK_TABLE_OR_COL Y) value)))))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-6 is a root stage , consists of Stage-7, Stage-8, Stage-1'
-'  Stage-7 has a backup stage: Stage-1'
-'  Stage-4 depends on stages: Stage-7'
-'  Stage-2 depends on stages: Stage-1, Stage-4, Stage-5'
-'  Stage-8 has a backup stage: Stage-1'
-'  Stage-5 depends on stages: Stage-8'
-'  Stage-1'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-6'
-'    Conditional Operator'
-''
-'  Stage: Stage-7'
-'    Map Reduce Local Work'
-'      Alias -> Map Local Tables:'
-'        y:src '
-'          Fetch Operator'
-'            limit: -1'
-'      Alias -> Map Local Operator Tree:'
-'        y:src '
-'          TableScan'
-'            alias: src'
-'            Select Operator'
-'              expressions:'
-'                    expr: key'
-'                    type: string'
-'                    expr: value'
-'                    type: string'
-'              outputColumnNames: _col0, _col1'
-'              HashTable Sink Operator'
-'                condition expressions:'
-'                  0 '
-'                  1 {_col0} {_col1}'
-'                handleSkewJoin: false'
-'                keys:'
-'                  0 [Column[_col0]]'
-'                  1 [Column[_col0]]'
-'                Position of Big Table: 0'
-''
-'  Stage: Stage-4'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        x:src '
-'          TableScan'
-'            alias: src'
-'            Select Operator'
-'              expressions:'
-'                    expr: key'
-'                    type: string'
-'              outputColumnNames: _col0'
-'              Map Join Operator'
-'                condition map:'
-'                     Inner Join 0 to 1'
-'                condition expressions:'
-'                  0 '
-'                  1 {_col0} {_col1}'
-'                handleSkewJoin: false'
-'                keys:'
-'                  0 [Column[_col0]]'
-'                  1 [Column[_col0]]'
-'                outputColumnNames: _col2, _col3'
-'                Position of Big Table: 0'
-'                Select Operator'
-'                  expressions:'
-'                        expr: _col2'
-'                        type: string'
-'                        expr: _col3'
-'                        type: string'
-'                  outputColumnNames: _col2, _col3'
-'                  Group By Operator'
-'                    aggregations:'
-'                          expr: sum(hash(_col2,_col3))'
-'                    bucketGroup: false'
-'                    mode: hash'
-'                    outputColumnNames: _col0'
-'                    File Output Operator'
-'                      compressed: false'
-'                      GlobalTableId: 0'
-'                      table:'
-'                          input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-'                          output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-'      Local Work:'
-'        Map Reduce Local Work'
-''
-'  Stage: Stage-2'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        file:!!{hive.exec.scratchdir}!! '
-'            Reduce Output Operator'
-'              sort order: '
-'              tag: -1'
-'              value expressions:'
-'                    expr: _col0'
-'                    type: bigint'
-'      Reduce Operator Tree:'
-'        Group By Operator'
-'          aggregations:'
-'                expr: sum(VALUE._col0)'
-'          bucketGroup: false'
-'          mode: mergepartial'
-'          outputColumnNames: _col0'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col0'
-'                  type: bigint'
-'            outputColumnNames: _col0'
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-8'
-'    Map Reduce Local Work'
-'      Alias -> Map Local Tables:'
-'        x:src '
-'          Fetch Operator'
-'            limit: -1'
-'      Alias -> Map Local Operator Tree:'
-'        x:src '
-'          TableScan'
-'            alias: src'
-'            Select Operator'
-'              expressions:'
-'                    expr: key'
-'                    type: string'
-'              outputColumnNames: _col0'
-'              HashTable Sink Operator'
-'                condition expressions:'
-'                  0 '
-'                  1 {_col0} {_col1}'
-'                handleSkewJoin: false'
-'                keys:'
-'                  0 [Column[_col0]]'
-'                  1 [Column[_col0]]'
-'                Position of Big Table: 1'
-''
-'  Stage: Stage-5'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        y:src '
-'          TableScan'
-'            alias: src'
-'            Select Operator'
-'              expressions:'
-'                    expr: key'
-'                    type: string'
-'                    expr: value'
-'                    type: string'
-'              outputColumnNames: _col0, _col1'
-'              Map Join Operator'
-'                condition map:'
-'                     Inner Join 0 to 1'
-'                condition expressions:'
-'                  0 '
-'                  1 {_col0} {_col1}'
-'                handleSkewJoin: false'
-'                keys:'
-'                  0 [Column[_col0]]'
-'                  1 [Column[_col0]]'
-'                outputColumnNames: _col2, _col3'
-'                Position of Big Table: 1'
-'                Select Operator'
-'                  expressions:'
-'                        expr: _col2'
-'                        type: string'
-'                        expr: _col3'
-'                        type: string'
-'                  outputColumnNames: _col2, _col3'
-'                  Group By Operator'
-'                    aggregations:'
-'                          expr: sum(hash(_col2,_col3))'
-'                    bucketGroup: false'
-'                    mode: hash'
-'                    outputColumnNames: _col0'
-'                    File Output Operator'
-'                      compressed: false'
-'                      GlobalTableId: 0'
-'                      table:'
-'                          input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-'                          output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-'      Local Work:'
-'        Map Reduce Local Work'
-''
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        x:src '
-'          TableScan'
-'            alias: src'
-'            Select Operator'
-'              expressions:'
-'                    expr: key'
-'                    type: string'
-'              outputColumnNames: _col0'
-'              Reduce Output Operator'
-'                key expressions:'
-'                      expr: _col0'
-'                      type: string'
-'                sort order: +'
-'                Map-reduce partition columns:'
-'                      expr: _col0'
-'                      type: string'
-'                tag: 0'
-'        y:src '
-'          TableScan'
-'            alias: src'
-'            Select Operator'
-'              expressions:'
-'                    expr: key'
-'                    type: string'
-'                    expr: value'
-'                    type: string'
-'              outputColumnNames: _col0, _col1'
-'              Reduce Output Operator'
-'                key expressions:'
-'                      expr: _col0'
-'                      type: string'
-'                sort order: +'
-'                Map-reduce partition columns:'
-'                      expr: _col0'
-'                      type: string'
-'                tag: 1'
-'                value expressions:'
-'                      expr: _col0'
-'                      type: string'
-'                      expr: _col1'
-'                      type: string'
-'      Reduce Operator Tree:'
-'        Join Operator'
-'          condition map:'
-'               Inner Join 0 to 1'
-'          condition expressions:'
-'            0 '
-'            1 {VALUE._col0} {VALUE._col1}'
-'          handleSkewJoin: false'
-'          outputColumnNames: _col2, _col3'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col2'
-'                  type: string'
-'                  expr: _col3'
-'                  type: string'
-'            outputColumnNames: _col2, _col3'
-'            Group By Operator'
-'              aggregations:'
-'                    expr: sum(hash(_col2,_col3))'
-'              bucketGroup: false'
-'              mode: hash'
-'              outputColumnNames: _col0'
-'              File Output Operator'
-'                compressed: false'
-'                GlobalTableId: 0'
-'                table:'
-'                    input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-'                    output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-''
-''
-268 rows selected 
->>>  
->>>  FROM 
-(SELECT src.* FROM src) x 
-JOIN 
-(SELECT src.* FROM src) Y 
-ON (x.key = Y.key) 
-select sum(hash(Y.key,Y.value));
-'_c0'
-'103231310608'
-1 row selected 
->>>  
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/auto_join11.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/auto_join11.q.out b/ql/src/test/results/beelinepositive/auto_join11.q.out
deleted file mode 100644
index 839d70b..0000000
--- a/ql/src/test/results/beelinepositive/auto_join11.q.out
+++ /dev/null
@@ -1,318 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/auto_join11.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/auto_join11.q
->>>  set hive.auto.convert.join = true;
-No rows affected 
->>>  
->>>  explain 
-SELECT sum(hash(src1.c1, src2.c4)) 
-FROM 
-(SELECT src.key as c1, src.value as c2 from src) src1 
-JOIN 
-(SELECT src.key as c3, src.value as c4 from src) src2 
-ON src1.c1 = src2.c3 AND src1.c1 < 100;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) key) c1) (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) value) c2)))) src1) (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) key) c3) (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) value) c4)))) src2) (AND (= (. (TOK_TABLE_OR_COL src1) c1) (. (TOK_TABLE_OR_COL src2) c3)) (< (. (TOK_TABLE_OR_COL src1) c1) 100)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION sum (TOK_FUNCTION hash (. (TOK_TABLE_OR_COL src1) c1) (. (TOK_TABLE_OR_COL src2) c4)))))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-6 is a root stage , consists of Stage-7, Stage-8, Stage-1'
-'  Stage-7 has a backup stage: Stage-1'
-'  Stage-4 depends on stages: Stage-7'
-'  Stage-2 depends on stages: Stage-1, Stage-4, Stage-5'
-'  Stage-8 has a backup stage: Stage-1'
-'  Stage-5 depends on stages: Stage-8'
-'  Stage-1'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-6'
-'    Conditional Operator'
-''
-'  Stage: Stage-7'
-'    Map Reduce Local Work'
-'      Alias -> Map Local Tables:'
-'        src2:src '
-'          Fetch Operator'
-'            limit: -1'
-'      Alias -> Map Local Operator Tree:'
-'        src2:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key < 100.0)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                HashTable Sink Operator'
-'                  condition expressions:'
-'                    0 {_col0}'
-'                    1 {_col1}'
-'                  handleSkewJoin: false'
-'                  keys:'
-'                    0 [Column[_col0]]'
-'                    1 [Column[_col0]]'
-'                  Position of Big Table: 0'
-''
-'  Stage: Stage-4'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src1:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key < 100.0)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                outputColumnNames: _col0'
-'                Map Join Operator'
-'                  condition map:'
-'                       Inner Join 0 to 1'
-'                  condition expressions:'
-'                    0 {_col0}'
-'                    1 {_col1}'
-'                  handleSkewJoin: false'
-'                  keys:'
-'                    0 [Column[_col0]]'
-'                    1 [Column[_col0]]'
-'                  outputColumnNames: _col0, _col3'
-'                  Position of Big Table: 0'
-'                  Select Operator'
-'                    expressions:'
-'                          expr: _col0'
-'                          type: string'
-'                          expr: _col3'
-'                          type: string'
-'                    outputColumnNames: _col0, _col3'
-'                    Group By Operator'
-'                      aggregations:'
-'                            expr: sum(hash(_col0,_col3))'
-'                      bucketGroup: false'
-'                      mode: hash'
-'                      outputColumnNames: _col0'
-'                      File Output Operator'
-'                        compressed: false'
-'                        GlobalTableId: 0'
-'                        table:'
-'                            input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-'                            output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-'      Local Work:'
-'        Map Reduce Local Work'
-''
-'  Stage: Stage-2'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        file:!!{hive.exec.scratchdir}!! '
-'            Reduce Output Operator'
-'              sort order: '
-'              tag: -1'
-'              value expressions:'
-'                    expr: _col0'
-'                    type: bigint'
-'      Reduce Operator Tree:'
-'        Group By Operator'
-'          aggregations:'
-'                expr: sum(VALUE._col0)'
-'          bucketGroup: false'
-'          mode: mergepartial'
-'          outputColumnNames: _col0'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col0'
-'                  type: bigint'
-'            outputColumnNames: _col0'
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-8'
-'    Map Reduce Local Work'
-'      Alias -> Map Local Tables:'
-'        src1:src '
-'          Fetch Operator'
-'            limit: -1'
-'      Alias -> Map Local Operator Tree:'
-'        src1:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key < 100.0)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                outputColumnNames: _col0'
-'                HashTable Sink Operator'
-'                  condition expressions:'
-'                    0 {_col0}'
-'                    1 {_col1}'
-'                  handleSkewJoin: false'
-'                  keys:'
-'                    0 [Column[_col0]]'
-'                    1 [Column[_col0]]'
-'                  Position of Big Table: 1'
-''
-'  Stage: Stage-5'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src2:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key < 100.0)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                Map Join Operator'
-'                  condition map:'
-'                       Inner Join 0 to 1'
-'                  condition expressions:'
-'                    0 {_col0}'
-'                    1 {_col1}'
-'                  handleSkewJoin: false'
-'                  keys:'
-'                    0 [Column[_col0]]'
-'                    1 [Column[_col0]]'
-'                  outputColumnNames: _col0, _col3'
-'                  Position of Big Table: 1'
-'                  Select Operator'
-'                    expressions:'
-'                          expr: _col0'
-'                          type: string'
-'                          expr: _col3'
-'                          type: string'
-'                    outputColumnNames: _col0, _col3'
-'                    Group By Operator'
-'                      aggregations:'
-'                            expr: sum(hash(_col0,_col3))'
-'                      bucketGroup: false'
-'                      mode: hash'
-'                      outputColumnNames: _col0'
-'                      File Output Operator'
-'                        compressed: false'
-'                        GlobalTableId: 0'
-'                        table:'
-'                            input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-'                            output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-'      Local Work:'
-'        Map Reduce Local Work'
-''
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src1:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key < 100.0)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                outputColumnNames: _col0'
-'                Reduce Output Operator'
-'                  key expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                  sort order: +'
-'                  Map-reduce partition columns:'
-'                        expr: _col0'
-'                        type: string'
-'                  tag: 0'
-'                  value expressions:'
-'                        expr: _col0'
-'                        type: string'
-'        src2:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key < 100.0)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                Reduce Output Operator'
-'                  key expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                  sort order: +'
-'                  Map-reduce partition columns:'
-'                        expr: _col0'
-'                        type: string'
-'                  tag: 1'
-'                  value expressions:'
-'                        expr: _col1'
-'                        type: string'
-'      Reduce Operator Tree:'
-'        Join Operator'
-'          condition map:'
-'               Inner Join 0 to 1'
-'          condition expressions:'
-'            0 {VALUE._col0}'
-'            1 {VALUE._col1}'
-'          handleSkewJoin: false'
-'          outputColumnNames: _col0, _col3'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col0'
-'                  type: string'
-'                  expr: _col3'
-'                  type: string'
-'            outputColumnNames: _col0, _col3'
-'            Group By Operator'
-'              aggregations:'
-'                    expr: sum(hash(_col0,_col3))'
-'              bucketGroup: false'
-'              mode: hash'
-'              outputColumnNames: _col0'
-'              File Output Operator'
-'                compressed: false'
-'                GlobalTableId: 0'
-'                table:'
-'                    input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-'                    output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-''
-''
-293 rows selected 
->>>  
->>>  SELECT sum(hash(src1.c1, src2.c4)) 
-FROM 
-(SELECT src.key as c1, src.value as c2 from src) src1 
-JOIN 
-(SELECT src.key as c3, src.value as c4 from src) src2 
-ON src1.c1 = src2.c3 AND src1.c1 < 100;
-'_c0'
-'-101333194320'
-1 row selected 
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/auto_join12.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/auto_join12.q.out b/ql/src/test/results/beelinepositive/auto_join12.q.out
deleted file mode 100644
index 7083a19..0000000
--- a/ql/src/test/results/beelinepositive/auto_join12.q.out
+++ /dev/null
@@ -1,525 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/auto_join12.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/auto_join12.q
->>>  
->>>  
->>>  set hive.auto.convert.join = true;
-No rows affected 
->>>  
->>>  
->>>  explain 
-SELECT sum(hash(src1.c1, src2.c4)) 
-FROM 
-(SELECT src.key as c1, src.value as c2 from src) src1 
-JOIN 
-(SELECT src.key as c3, src.value as c4 from src) src2 
-ON src1.c1 = src2.c3 AND src1.c1 < 100 
-JOIN 
-(SELECT src.key as c5, src.value as c6 from src) src3 
-ON src1.c1 = src3.c5 AND src3.c5 < 80;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_JOIN (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) key) c1) (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) value) c2)))) src1) (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) key) c3) (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) value) c4)))) src2) (AND (= (. (TOK_TABLE_OR_COL src1) c1) (. (TOK_TABLE_OR_COL src2) c3)) (< (. (TOK_TABLE_OR_COL src1) c1) 100))) (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) key) c5) (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) value) c6)))) src3) (AND (= (. (TOK_TABLE_OR_COL src1) c1) (. (TOK_TABLE_OR_COL src3) c5)) (< (. (TOK_TABLE_OR_COL src3) c5) 80)))) (TOK_INSERT 
 (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION sum (TOK_FUNCTION hash (. (TOK_TABLE_OR_COL src1) c1) (. (TOK_TABLE_OR_COL src2) c4)))))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-8 is a root stage , consists of Stage-9, Stage-10, Stage-11, Stage-1'
-'  Stage-9 has a backup stage: Stage-1'
-'  Stage-5 depends on stages: Stage-9'
-'  Stage-2 depends on stages: Stage-1, Stage-5, Stage-6, Stage-7'
-'  Stage-10 has a backup stage: Stage-1'
-'  Stage-6 depends on stages: Stage-10'
-'  Stage-11 has a backup stage: Stage-1'
-'  Stage-7 depends on stages: Stage-11'
-'  Stage-1'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-8'
-'    Conditional Operator'
-''
-'  Stage: Stage-9'
-'    Map Reduce Local Work'
-'      Alias -> Map Local Tables:'
-'        src2:src '
-'          Fetch Operator'
-'            limit: -1'
-'        src3:src '
-'          Fetch Operator'
-'            limit: -1'
-'      Alias -> Map Local Operator Tree:'
-'        src2:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key < 100) and (key < 80))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                HashTable Sink Operator'
-'                  condition expressions:'
-'                    0 {_col0}'
-'                    1 {_col1}'
-'                    2 '
-'                  handleSkewJoin: false'
-'                  keys:'
-'                    0 [Column[_col0]]'
-'                    1 [Column[_col0]]'
-'                    2 [Column[_col0]]'
-'                  Position of Big Table: 0'
-'        src3:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key < 80) and (key < 100))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                outputColumnNames: _col0'
-'                HashTable Sink Operator'
-'                  condition expressions:'
-'                    0 {_col0}'
-'                    1 {_col1}'
-'                    2 '
-'                  handleSkewJoin: false'
-'                  keys:'
-'                    0 [Column[_col0]]'
-'                    1 [Column[_col0]]'
-'                    2 [Column[_col0]]'
-'                  Position of Big Table: 0'
-''
-'  Stage: Stage-5'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src1:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key < 100) and (key < 80))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                outputColumnNames: _col0'
-'                Map Join Operator'
-'                  condition map:'
-'                       Inner Join 0 to 1'
-'                       Inner Join 0 to 2'
-'                  condition expressions:'
-'                    0 {_col0}'
-'                    1 {_col1}'
-'                    2 '
-'                  handleSkewJoin: false'
-'                  keys:'
-'                    0 [Column[_col0]]'
-'                    1 [Column[_col0]]'
-'                    2 [Column[_col0]]'
-'                  outputColumnNames: _col0, _col3'
-'                  Position of Big Table: 0'
-'                  Select Operator'
-'                    expressions:'
-'                          expr: _col0'
-'                          type: string'
-'                          expr: _col3'
-'                          type: string'
-'                    outputColumnNames: _col0, _col3'
-'                    Group By Operator'
-'                      aggregations:'
-'                            expr: sum(hash(_col0,_col3))'
-'                      bucketGroup: false'
-'                      mode: hash'
-'                      outputColumnNames: _col0'
-'                      File Output Operator'
-'                        compressed: false'
-'                        GlobalTableId: 0'
-'                        table:'
-'                            input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-'                            output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-'      Local Work:'
-'        Map Reduce Local Work'
-''
-'  Stage: Stage-2'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        file:!!{hive.exec.scratchdir}!! '
-'            Reduce Output Operator'
-'              sort order: '
-'              tag: -1'
-'              value expressions:'
-'                    expr: _col0'
-'                    type: bigint'
-'      Reduce Operator Tree:'
-'        Group By Operator'
-'          aggregations:'
-'                expr: sum(VALUE._col0)'
-'          bucketGroup: false'
-'          mode: mergepartial'
-'          outputColumnNames: _col0'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col0'
-'                  type: bigint'
-'            outputColumnNames: _col0'
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-10'
-'    Map Reduce Local Work'
-'      Alias -> Map Local Tables:'
-'        src1:src '
-'          Fetch Operator'
-'            limit: -1'
-'        src3:src '
-'          Fetch Operator'
-'            limit: -1'
-'      Alias -> Map Local Operator Tree:'
-'        src1:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key < 100) and (key < 80))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                outputColumnNames: _col0'
-'                HashTable Sink Operator'
-'                  condition expressions:'
-'                    0 {_col0}'
-'                    1 {_col1}'
-'                    2 '
-'                  handleSkewJoin: false'
-'                  keys:'
-'                    0 [Column[_col0]]'
-'                    1 [Column[_col0]]'
-'                    2 [Column[_col0]]'
-'                  Position of Big Table: 1'
-'        src3:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key < 80) and (key < 100))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                outputColumnNames: _col0'
-'                HashTable Sink Operator'
-'                  condition expressions:'
-'                    0 {_col0}'
-'                    1 {_col1}'
-'                    2 '
-'                  handleSkewJoin: false'
-'                  keys:'
-'                    0 [Column[_col0]]'
-'                    1 [Column[_col0]]'
-'                    2 [Column[_col0]]'
-'                  Position of Big Table: 1'
-''
-'  Stage: Stage-6'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src2:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key < 100) and (key < 80))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                Map Join Operator'
-'                  condition map:'
-'                       Inner Join 0 to 1'
-'                       Inner Join 0 to 2'
-'                  condition expressions:'
-'                    0 {_col0}'
-'                    1 {_col1}'
-'                    2 '
-'                  handleSkewJoin: false'
-'                  keys:'
-'                    0 [Column[_col0]]'
-'                    1 [Column[_col0]]'
-'                    2 [Column[_col0]]'
-'                  outputColumnNames: _col0, _col3'
-'                  Position of Big Table: 1'
-'                  Select Operator'
-'                    expressions:'
-'                          expr: _col0'
-'                          type: string'
-'                          expr: _col3'
-'                          type: string'
-'                    outputColumnNames: _col0, _col3'
-'                    Group By Operator'
-'                      aggregations:'
-'                            expr: sum(hash(_col0,_col3))'
-'                      bucketGroup: false'
-'                      mode: hash'
-'                      outputColumnNames: _col0'
-'                      File Output Operator'
-'                        compressed: false'
-'                        GlobalTableId: 0'
-'                        table:'
-'                            input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-'                            output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-'      Local Work:'
-'        Map Reduce Local Work'
-''
-'  Stage: Stage-11'
-'    Map Reduce Local Work'
-'      Alias -> Map Local Tables:'
-'        src1:src '
-'          Fetch Operator'
-'            limit: -1'
-'        src2:src '
-'          Fetch Operator'
-'            limit: -1'
-'      Alias -> Map Local Operator Tree:'
-'        src1:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key < 100) and (key < 80))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                outputColumnNames: _col0'
-'                HashTable Sink Operator'
-'                  condition expressions:'
-'                    0 {_col0}'
-'                    1 {_col1}'
-'                    2 '
-'                  handleSkewJoin: false'
-'                  keys:'
-'                    0 [Column[_col0]]'
-'                    1 [Column[_col0]]'
-'                    2 [Column[_col0]]'
-'                  Position of Big Table: 2'
-'        src2:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key < 100) and (key < 80))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                HashTable Sink Operator'
-'                  condition expressions:'
-'                    0 {_col0}'
-'                    1 {_col1}'
-'                    2 '
-'                  handleSkewJoin: false'
-'                  keys:'
-'                    0 [Column[_col0]]'
-'                    1 [Column[_col0]]'
-'                    2 [Column[_col0]]'
-'                  Position of Big Table: 2'
-''
-'  Stage: Stage-7'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src3:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key < 80) and (key < 100))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                outputColumnNames: _col0'
-'                Map Join Operator'
-'                  condition map:'
-'                       Inner Join 0 to 1'
-'                       Inner Join 0 to 2'
-'                  condition expressions:'
-'                    0 {_col0}'
-'                    1 {_col1}'
-'                    2 '
-'                  handleSkewJoin: false'
-'                  keys:'
-'                    0 [Column[_col0]]'
-'                    1 [Column[_col0]]'
-'                    2 [Column[_col0]]'
-'                  outputColumnNames: _col0, _col3'
-'                  Position of Big Table: 2'
-'                  Select Operator'
-'                    expressions:'
-'                          expr: _col0'
-'                          type: string'
-'                          expr: _col3'
-'                          type: string'
-'                    outputColumnNames: _col0, _col3'
-'                    Group By Operator'
-'                      aggregations:'
-'                            expr: sum(hash(_col0,_col3))'
-'                      bucketGroup: false'
-'                      mode: hash'
-'                      outputColumnNames: _col0'
-'                      File Output Operator'
-'                        compressed: false'
-'                        GlobalTableId: 0'
-'                        table:'
-'                            input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-'                            output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-'      Local Work:'
-'        Map Reduce Local Work'
-''
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src1:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key < 100) and (key < 80))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                outputColumnNames: _col0'
-'                Reduce Output Operator'
-'                  key expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                  sort order: +'
-'                  Map-reduce partition columns:'
-'                        expr: _col0'
-'                        type: string'
-'                  tag: 0'
-'                  value expressions:'
-'                        expr: _col0'
-'                        type: string'
-'        src2:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key < 100) and (key < 80))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                Reduce Output Operator'
-'                  key expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                  sort order: +'
-'                  Map-reduce partition columns:'
-'                        expr: _col0'
-'                        type: string'
-'                  tag: 1'
-'                  value expressions:'
-'                        expr: _col1'
-'                        type: string'
-'        src3:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key < 80) and (key < 100))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                outputColumnNames: _col0'
-'                Reduce Output Operator'
-'                  key expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                  sort order: +'
-'                  Map-reduce partition columns:'
-'                        expr: _col0'
-'                        type: string'
-'                  tag: 2'
-'      Reduce Operator Tree:'
-'        Join Operator'
-'          condition map:'
-'               Inner Join 0 to 1'
-'               Inner Join 0 to 2'
-'          condition expressions:'
-'            0 {VALUE._col0}'
-'            1 {VALUE._col1}'
-'            2 '
-'          handleSkewJoin: false'
-'          outputColumnNames: _col0, _col3'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col0'
-'                  type: string'
-'                  expr: _col3'
-'                  type: string'
-'            outputColumnNames: _col0, _col3'
-'            Group By Operator'
-'              aggregations:'
-'                    expr: sum(hash(_col0,_col3))'
-'              bucketGroup: false'
-'              mode: hash'
-'              outputColumnNames: _col0'
-'              File Output Operator'
-'                compressed: false'
-'                GlobalTableId: 0'
-'                table:'
-'                    input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-'                    output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-''
-''
-491 rows selected 
->>>  
->>>  SELECT sum(hash(src1.c1, src2.c4)) 
-FROM 
-(SELECT src.key as c1, src.value as c2 from src) src1 
-JOIN 
-(SELECT src.key as c3, src.value as c4 from src) src2 
-ON src1.c1 = src2.c3 AND src1.c1 < 100 
-JOIN 
-(SELECT src.key as c5, src.value as c6 from src) src3 
-ON src1.c1 = src3.c5 AND src3.c5 < 80;
-'_c0'
-'-136843922952'
-1 row selected 
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/auto_join13.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/auto_join13.q.out b/ql/src/test/results/beelinepositive/auto_join13.q.out
deleted file mode 100644
index 627b2c3..0000000
--- a/ql/src/test/results/beelinepositive/auto_join13.q.out
+++ /dev/null
@@ -1,499 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/auto_join13.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/auto_join13.q
->>>  
->>>  set hive.auto.convert.join = true;
-No rows affected 
->>>  
->>>  explain 
-SELECT sum(hash(src1.c1, src2.c4)) 
-FROM 
-(SELECT src.key as c1, src.value as c2 from src) src1 
-JOIN 
-(SELECT src.key as c3, src.value as c4 from src) src2 
-ON src1.c1 = src2.c3 AND src1.c1 < 100 
-JOIN 
-(SELECT src.key as c5, src.value as c6 from src) src3 
-ON src1.c1 + src2.c3 = src3.c5 AND src3.c5 < 200;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_JOIN (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) key) c1) (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) value) c2)))) src1) (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) key) c3) (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) value) c4)))) src2) (AND (= (. (TOK_TABLE_OR_COL src1) c1) (. (TOK_TABLE_OR_COL src2) c3)) (< (. (TOK_TABLE_OR_COL src1) c1) 100))) (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) key) c5) (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) value) c6)))) src3) (AND (= (+ (. (TOK_TABLE_OR_COL src1) c1) (. (TOK_TABLE_OR_COL src2) c3)) (. (TOK_TABLE_OR_COL src3) c5)) (< (. (TOK_TABLE_O
 R_COL src3) c5) 200)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION sum (TOK_FUNCTION hash (. (TOK_TABLE_OR_COL src1) c1) (. (TOK_TABLE_OR_COL src2) c4)))))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-11 is a root stage , consists of Stage-14, Stage-15, Stage-1'
-'  Stage-14 has a backup stage: Stage-1'
-'  Stage-9 depends on stages: Stage-14'
-'  Stage-8 depends on stages: Stage-1, Stage-9, Stage-10 , consists of Stage-12, Stage-13, Stage-2'
-'  Stage-12 has a backup stage: Stage-2'
-'  Stage-6 depends on stages: Stage-12'
-'  Stage-3 depends on stages: Stage-2, Stage-6, Stage-7'
-'  Stage-13 has a backup stage: Stage-2'
-'  Stage-7 depends on stages: Stage-13'
-'  Stage-2'
-'  Stage-15 has a backup stage: Stage-1'
-'  Stage-10 depends on stages: Stage-15'
-'  Stage-1'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-11'
-'    Conditional Operator'
-''
-'  Stage: Stage-14'
-'    Map Reduce Local Work'
-'      Alias -> Map Local Tables:'
-'        src2:src '
-'          Fetch Operator'
-'            limit: -1'
-'      Alias -> Map Local Operator Tree:'
-'        src2:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key < 100)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                HashTable Sink Operator'
-'                  condition expressions:'
-'                    0 {_col0}'
-'                    1 {_col0} {_col1}'
-'                  handleSkewJoin: false'
-'                  keys:'
-'                    0 [Column[_col0]]'
-'                    1 [Column[_col0]]'
-'                  Position of Big Table: 0'
-''
-'  Stage: Stage-9'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src1:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key < 100)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                outputColumnNames: _col0'
-'                Map Join Operator'
-'                  condition map:'
-'                       Inner Join 0 to 1'
-'                  condition expressions:'
-'                    0 {_col0}'
-'                    1 {_col0} {_col1}'
-'                  handleSkewJoin: false'
-'                  keys:'
-'                    0 [Column[_col0]]'
-'                    1 [Column[_col0]]'
-'                  outputColumnNames: _col0, _col2, _col3'
-'                  Position of Big Table: 0'
-'                  File Output Operator'
-'                    compressed: false'
-'                    GlobalTableId: 0'
-'                    table:'
-'                        input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-'                        output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-'      Local Work:'
-'        Map Reduce Local Work'
-''
-'  Stage: Stage-8'
-'    Conditional Operator'
-''
-'  Stage: Stage-12'
-'    Map Reduce Local Work'
-'      Alias -> Map Local Tables:'
-'        src3:src '
-'          Fetch Operator'
-'            limit: -1'
-'      Alias -> Map Local Operator Tree:'
-'        src3:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key < 200)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                outputColumnNames: _col0'
-'                HashTable Sink Operator'
-'                  condition expressions:'
-'                    0 {_col3} {_col0}'
-'                    1 '
-'                  handleSkewJoin: false'
-'                  keys:'
-'                    0 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0], Column[_col2]()]'
-'                    1 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0]()]'
-'                  Position of Big Table: 0'
-''
-'  Stage: Stage-6'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        $INTNAME '
-'            Map Join Operator'
-'              condition map:'
-'                   Inner Join 0 to 1'
-'              condition expressions:'
-'                0 {_col3} {_col0}'
-'                1 '
-'              handleSkewJoin: false'
-'              keys:'
-'                0 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0], Column[_col2]()]'
-'                1 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0]()]'
-'              outputColumnNames: _col1, _col2'
-'              Position of Big Table: 0'
-'              Select Operator'
-'                expressions:'
-'                      expr: _col2'
-'                      type: string'
-'                      expr: _col1'
-'                      type: string'
-'                outputColumnNames: _col2, _col1'
-'                Group By Operator'
-'                  aggregations:'
-'                        expr: sum(hash(_col2,_col1))'
-'                  bucketGroup: false'
-'                  mode: hash'
-'                  outputColumnNames: _col0'
-'                  File Output Operator'
-'                    compressed: false'
-'                    GlobalTableId: 0'
-'                    table:'
-'                        input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-'                        output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-'      Local Work:'
-'        Map Reduce Local Work'
-''
-'  Stage: Stage-3'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        file:!!{hive.exec.scratchdir}!! '
-'            Reduce Output Operator'
-'              sort order: '
-'              tag: -1'
-'              value expressions:'
-'                    expr: _col0'
-'                    type: bigint'
-'      Reduce Operator Tree:'
-'        Group By Operator'
-'          aggregations:'
-'                expr: sum(VALUE._col0)'
-'          bucketGroup: false'
-'          mode: mergepartial'
-'          outputColumnNames: _col0'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col0'
-'                  type: bigint'
-'            outputColumnNames: _col0'
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-13'
-'    Map Reduce Local Work'
-'      Alias -> Map Local Tables:'
-'        $INTNAME '
-'          Fetch Operator'
-'            limit: -1'
-'      Alias -> Map Local Operator Tree:'
-'        $INTNAME '
-'            HashTable Sink Operator'
-'              condition expressions:'
-'                0 {_col3} {_col0}'
-'                1 '
-'              handleSkewJoin: false'
-'              keys:'
-'                0 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0], Column[_col2]()]'
-'                1 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0]()]'
-'              Position of Big Table: 1'
-''
-'  Stage: Stage-7'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src3:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key < 200)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                outputColumnNames: _col0'
-'                Map Join Operator'
-'                  condition map:'
-'                       Inner Join 0 to 1'
-'                  condition expressions:'
-'                    0 {_col3} {_col0}'
-'                    1 '
-'                  handleSkewJoin: false'
-'                  keys:'
-'                    0 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0], Column[_col2]()]'
-'                    1 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0]()]'
-'                  outputColumnNames: _col1, _col2'
-'                  Position of Big Table: 1'
-'                  Select Operator'
-'                    expressions:'
-'                          expr: _col2'
-'                          type: string'
-'                          expr: _col1'
-'                          type: string'
-'                    outputColumnNames: _col2, _col1'
-'                    Group By Operator'
-'                      aggregations:'
-'                            expr: sum(hash(_col2,_col1))'
-'                      bucketGroup: false'
-'                      mode: hash'
-'                      outputColumnNames: _col0'
-'                      File Output Operator'
-'                        compressed: false'
-'                        GlobalTableId: 0'
-'                        table:'
-'                            input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-'                            output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-'      Local Work:'
-'        Map Reduce Local Work'
-''
-'  Stage: Stage-2'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        $INTNAME '
-'            Reduce Output Operator'
-'              key expressions:'
-'                    expr: (_col0 + _col2)'
-'                    type: double'
-'              sort order: +'
-'              Map-reduce partition columns:'
-'                    expr: (_col0 + _col2)'
-'                    type: double'
-'              tag: 0'
-'              value expressions:'
-'                    expr: _col3'
-'                    type: string'
-'                    expr: _col0'
-'                    type: string'
-'        src3:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key < 200)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                outputColumnNames: _col0'
-'                Reduce Output Operator'
-'                  key expressions:'
-'                        expr: UDFToDouble(_col0)'
-'                        type: double'
-'                  sort order: +'
-'                  Map-reduce partition columns:'
-'                        expr: UDFToDouble(_col0)'
-'                        type: double'
-'                  tag: 1'
-'      Reduce Operator Tree:'
-'        Join Operator'
-'          condition map:'
-'               Inner Join 0 to 1'
-'          condition expressions:'
-'            0 {VALUE._col1} {VALUE._col2}'
-'            1 '
-'          handleSkewJoin: false'
-'          outputColumnNames: _col1, _col2'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col2'
-'                  type: string'
-'                  expr: _col1'
-'                  type: string'
-'            outputColumnNames: _col2, _col1'
-'            Group By Operator'
-'              aggregations:'
-'                    expr: sum(hash(_col2,_col1))'
-'              bucketGroup: false'
-'              mode: hash'
-'              outputColumnNames: _col0'
-'              File Output Operator'
-'                compressed: false'
-'                GlobalTableId: 0'
-'                table:'
-'                    input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-'                    output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-''
-'  Stage: Stage-15'
-'    Map Reduce Local Work'
-'      Alias -> Map Local Tables:'
-'        src1:src '
-'          Fetch Operator'
-'            limit: -1'
-'      Alias -> Map Local Operator Tree:'
-'        src1:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key < 100)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                outputColumnNames: _col0'
-'                HashTable Sink Operator'
-'                  condition expressions:'
-'                    0 {_col0}'
-'                    1 {_col0} {_col1}'
-'                  handleSkewJoin: false'
-'                  keys:'
-'                    0 [Column[_col0]]'
-'                    1 [Column[_col0]]'
-'                  Position of Big Table: 1'
-''
-'  Stage: Stage-10'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src2:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key < 100)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                Map Join Operator'
-'                  condition map:'
-'                       Inner Join 0 to 1'
-'                  condition expressions:'
-'                    0 {_col0}'
-'                    1 {_col0} {_col1}'
-'                  handleSkewJoin: false'
-'                  keys:'
-'                    0 [Column[_col0]]'
-'                    1 [Column[_col0]]'
-'                  outputColumnNames: _col0, _col2, _col3'
-'                  Position of Big Table: 1'
-'                  File Output Operator'
-'                    compressed: false'
-'                    GlobalTableId: 0'
-'                    table:'
-'                        input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-'                        output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-'      Local Work:'
-'        Map Reduce Local Work'
-''
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src1:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key < 100)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                outputColumnNames: _col0'
-'                Reduce Output Operator'
-'                  key expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                  sort order: +'
-'                  Map-reduce partition columns:'
-'                        expr: _col0'
-'                        type: string'
-'                  tag: 0'
-'                  value expressions:'
-'                        expr: _col0'
-'                        type: string'
-'        src2:src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key < 100)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                Reduce Output Operator'
-'                  key expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                  sort order: +'
-'                  Map-reduce partition columns:'
-'                        expr: _col0'
-'                        type: string'
-'                  tag: 1'
-'                  value expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                        expr: _col1'
-'                        type: string'
-'      Reduce Operator Tree:'
-'        Join Operator'
-'          condition map:'
-'               Inner Join 0 to 1'
-'          condition expressions:'
-'            0 {VALUE._col0}'
-'            1 {VALUE._col0} {VALUE._col1}'
-'          handleSkewJoin: false'
-'          outputColumnNames: _col0, _col2, _col3'
-'          File Output Operator'
-'            compressed: false'
-'            GlobalTableId: 0'
-'            table:'
-'                input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-'                output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-''
-''
-467 rows selected 
->>>  
->>>  SELECT sum(hash(src1.c1, src2.c4)) 
-FROM 
-(SELECT src.key as c1, src.value as c2 from src) src1 
-JOIN 
-(SELECT src.key as c3, src.value as c4 from src) src2 
-ON src1.c1 = src2.c3 AND src1.c1 < 100 
-JOIN 
-(SELECT src.key as c5, src.value as c6 from src) src3 
-ON src1.c1 + src2.c3 = src3.c5 AND src3.c5 < 200;
-'_c0'
-'-97670109576'
-1 row selected 
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/auto_join14.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/auto_join14.q.out b/ql/src/test/results/beelinepositive/auto_join14.q.out
deleted file mode 100644
index 46d9ca5..0000000
--- a/ql/src/test/results/beelinepositive/auto_join14.q.out
+++ /dev/null
@@ -1,276 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/auto_join14.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/auto_join14.q
->>>  
->>>  set hive.auto.convert.join = true;
-No rows affected 
->>>  
->>>  CREATE TABLE dest1(c1 INT, c2 STRING) STORED AS TEXTFILE;
-No rows affected 
->>>  
->>>  set mapred.job.tracker=does.notexist.com:666;
-No rows affected 
->>>  set hive.exec.mode.local.auto=true;
-No rows affected 
->>>  
->>>  explain 
-FROM src JOIN srcpart ON src.key = srcpart.key AND srcpart.ds = '2008-04-08' and src.key > 100 
-INSERT OVERWRITE TABLE dest1 SELECT src.key, srcpart.value;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_TABREF (TOK_TABNAME src)) (TOK_TABREF (TOK_TABNAME srcpart)) (and (AND (= (. (TOK_TABLE_OR_COL src) key) (. (TOK_TABLE_OR_COL srcpart) key)) (= (. (TOK_TABLE_OR_COL srcpart) ds) '2008-04-08')) (> (. (TOK_TABLE_OR_COL src) key) 100)))) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME dest1))) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) key)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL srcpart) value)))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-6 is a root stage , consists of Stage-7, Stage-8, Stage-1'
-'  Stage-7 has a backup stage: Stage-1'
-'  Stage-4 depends on stages: Stage-7'
-'  Stage-0 depends on stages: Stage-1, Stage-4, Stage-5'
-'  Stage-2 depends on stages: Stage-0'
-'  Stage-8 has a backup stage: Stage-1'
-'  Stage-5 depends on stages: Stage-8'
-'  Stage-1'
-''
-'STAGE PLANS:'
-'  Stage: Stage-6'
-'    Conditional Operator'
-''
-'  Stage: Stage-7'
-'    Map Reduce Local Work'
-'      Alias -> Map Local Tables:'
-'        srcpart '
-'          Fetch Operator'
-'            limit: -1'
-'      Alias -> Map Local Operator Tree:'
-'        srcpart '
-'          TableScan'
-'            alias: srcpart'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key > 100.0)'
-'                  type: boolean'
-'              HashTable Sink Operator'
-'                condition expressions:'
-'                  0 {key}'
-'                  1 {value}'
-'                handleSkewJoin: false'
-'                keys:'
-'                  0 [Column[key]]'
-'                  1 [Column[key]]'
-'                Position of Big Table: 0'
-''
-'  Stage: Stage-4'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key > 100.0)'
-'                  type: boolean'
-'              Map Join Operator'
-'                condition map:'
-'                     Inner Join 0 to 1'
-'                condition expressions:'
-'                  0 {key}'
-'                  1 {value}'
-'                handleSkewJoin: false'
-'                keys:'
-'                  0 [Column[key]]'
-'                  1 [Column[key]]'
-'                outputColumnNames: _col0, _col5'
-'                Position of Big Table: 0'
-'                Select Operator'
-'                  expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                        expr: _col5'
-'                        type: string'
-'                  outputColumnNames: _col0, _col1'
-'                  Select Operator'
-'                    expressions:'
-'                          expr: UDFToInteger(_col0)'
-'                          type: int'
-'                          expr: _col1'
-'                          type: string'
-'                    outputColumnNames: _col0, _col1'
-'                    File Output Operator'
-'                      compressed: false'
-'                      GlobalTableId: 1'
-'                      table:'
-'                          input format: org.apache.hadoop.mapred.TextInputFormat'
-'                          output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                          serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                          name: auto_join14.dest1'
-'      Local Work:'
-'        Map Reduce Local Work'
-''
-'  Stage: Stage-0'
-'    Move Operator'
-'      tables:'
-'          replace: true'
-'          table:'
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: auto_join14.dest1'
-''
-'  Stage: Stage-2'
-'    Stats-Aggr Operator'
-''
-'  Stage: Stage-8'
-'    Map Reduce Local Work'
-'      Alias -> Map Local Tables:'
-'        src '
-'          Fetch Operator'
-'            limit: -1'
-'      Alias -> Map Local Operator Tree:'
-'        src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key > 100.0)'
-'                  type: boolean'
-'              HashTable Sink Operator'
-'                condition expressions:'
-'                  0 {key}'
-'                  1 {value}'
-'                handleSkewJoin: false'
-'                keys:'
-'                  0 [Column[key]]'
-'                  1 [Column[key]]'
-'                Position of Big Table: 1'
-''
-'  Stage: Stage-5'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        srcpart '
-'          TableScan'
-'            alias: srcpart'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key > 100.0)'
-'                  type: boolean'
-'              Map Join Operator'
-'                condition map:'
-'                     Inner Join 0 to 1'
-'                condition expressions:'
-'                  0 {key}'
-'                  1 {value}'
-'                handleSkewJoin: false'
-'                keys:'
-'                  0 [Column[key]]'
-'                  1 [Column[key]]'
-'                outputColumnNames: _col0, _col5'
-'                Position of Big Table: 1'
-'                Select Operator'
-'                  expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                        expr: _col5'
-'                        type: string'
-'                  outputColumnNames: _col0, _col1'
-'                  Select Operator'
-'                    expressions:'
-'                          expr: UDFToInteger(_col0)'
-'                          type: int'
-'                          expr: _col1'
-'                          type: string'
-'                    outputColumnNames: _col0, _col1'
-'                    File Output Operator'
-'                      compressed: false'
-'                      GlobalTableId: 1'
-'                      table:'
-'                          input format: org.apache.hadoop.mapred.TextInputFormat'
-'                          output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                          serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                          name: auto_join14.dest1'
-'      Local Work:'
-'        Map Reduce Local Work'
-''
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key > 100.0)'
-'                  type: boolean'
-'              Reduce Output Operator'
-'                key expressions:'
-'                      expr: key'
-'                      type: string'
-'                sort order: +'
-'                Map-reduce partition columns:'
-'                      expr: key'
-'                      type: string'
-'                tag: 0'
-'                value expressions:'
-'                      expr: key'
-'                      type: string'
-'        srcpart '
-'          TableScan'
-'            alias: srcpart'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key > 100.0)'
-'                  type: boolean'
-'              Reduce Output Operator'
-'                key expressions:'
-'                      expr: key'
-'                      type: string'
-'                sort order: +'
-'                Map-reduce partition columns:'
-'                      expr: key'
-'                      type: string'
-'                tag: 1'
-'                value expressions:'
-'                      expr: value'
-'                      type: string'
-'      Reduce Operator Tree:'
-'        Join Operator'
-'          condition map:'
-'               Inner Join 0 to 1'
-'          condition expressions:'
-'            0 {VALUE._col0}'
-'            1 {VALUE._col1}'
-'          handleSkewJoin: false'
-'          outputColumnNames: _col0, _col5'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col0'
-'                  type: string'
-'                  expr: _col5'
-'                  type: string'
-'            outputColumnNames: _col0, _col1'
-'            Select Operator'
-'              expressions:'
-'                    expr: UDFToInteger(_col0)'
-'                    type: int'
-'                    expr: _col1'
-'                    type: string'
-'              outputColumnNames: _col0, _col1'
-'              File Output Operator'
-'                compressed: false'
-'                GlobalTableId: 1'
-'                table:'
-'                    input format: org.apache.hadoop.mapred.TextInputFormat'
-'                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                    name: auto_join14.dest1'
-''
-''
-246 rows selected 
->>>  
->>>  FROM src JOIN srcpart ON src.key = srcpart.key AND srcpart.ds = '2008-04-08' and src.key > 100 
-INSERT OVERWRITE TABLE dest1 SELECT src.key, srcpart.value;
-'_col0','_col1'
-No rows selected 
->>>  
->>>  SELECT sum(hash(dest1.c1,dest1.c2)) FROM dest1;
-'_c0'
-'404554174174'
-1 row selected 
->>>  !record