You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by gu...@apache.org on 2017/02/03 21:50:23 UTC

[10/51] [partial] hive git commit: HIVE-15790: Remove unused beeline golden files (Gunther Hagleitner, reviewed by Sergey Shelukhin)

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/index_auto_partitioned.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/index_auto_partitioned.q.out b/ql/src/test/results/beelinepositive/index_auto_partitioned.q.out
deleted file mode 100644
index 1b7a226..0000000
--- a/ql/src/test/results/beelinepositive/index_auto_partitioned.q.out
+++ /dev/null
@@ -1,157 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/index_auto_partitioned.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/index_auto_partitioned.q
->>>  -- test automatic use of index on table with partitions
->>>  CREATE INDEX src_part_index ON TABLE srcpart(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected 
->>>  ALTER INDEX src_part_index ON srcpart REBUILD;
-No rows affected 
->>>  
->>>  SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected 
->>>  SET hive.optimize.index.filter=true;
-No rows affected 
->>>  SET hive.optimize.index.filter.compact.minsize=0;
-No rows affected 
->>>  
->>>  EXPLAIN SELECT key, value FROM srcpart WHERE key=86 AND ds='2008-04-09' ORDER BY key;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME srcpart))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_TABLE_OR_COL value))) (TOK_WHERE (AND (= (TOK_TABLE_OR_COL key) 86) (= (TOK_TABLE_OR_COL ds) '2008-04-09'))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-3 is a root stage'
-'  Stage-8 depends on stages: Stage-3 , consists of Stage-5, Stage-4, Stage-6'
-'  Stage-5'
-'  Stage-2 depends on stages: Stage-5, Stage-4, Stage-7'
-'  Stage-1 depends on stages: Stage-2'
-'  Stage-4'
-'  Stage-6'
-'  Stage-7 depends on stages: Stage-6'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-3'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        index_auto_partitioned__srcpart_src_part_index__ '
-'          TableScan'
-'            alias: index_auto_partitioned__srcpart_src_part_index__'
-'            filterExpr:'
-'                expr: ((key = 86.0) and (ds = '2008-04-09'))'
-'                type: boolean'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key = 86.0)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: _bucketname'
-'                      type: string'
-'                      expr: _offsets'
-'                      type: array<bigint>'
-'                outputColumnNames: _col0, _col1'
-'                File Output Operator'
-'                  compressed: false'
-'                  GlobalTableId: 1'
-'                  table:'
-'                      input format: org.apache.hadoop.mapred.TextInputFormat'
-'                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-8'
-'    Conditional Operator'
-''
-'  Stage: Stage-5'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          destination: file:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-2'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          destination: file:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        srcpart '
-'          TableScan'
-'            alias: srcpart'
-'            filterExpr:'
-'                expr: ((key = 86.0) and (ds = '2008-04-09'))'
-'                type: boolean'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key = 86.0)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                Reduce Output Operator'
-'                  key expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                  sort order: +'
-'                  tag: -1'
-'                  value expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                        expr: _col1'
-'                        type: string'
-'      Reduce Operator Tree:'
-'        Extract'
-'          File Output Operator'
-'            compressed: false'
-'            GlobalTableId: 0'
-'            table:'
-'                input format: org.apache.hadoop.mapred.TextInputFormat'
-'                output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-4'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        file:!!{hive.exec.scratchdir}!! '
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-6'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        file:!!{hive.exec.scratchdir}!! '
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-7'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          destination: file:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-''
-''
-130 rows selected 
->>>  SELECT key, value FROM srcpart WHERE key=86 AND ds='2008-04-09' ORDER BY key;
-'key','value'
-'86','val_86'
-'86','val_86'
-2 rows selected 
->>>  
->>>  DROP INDEX src_part_index ON srcpart;
-No rows affected 
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/index_auto_self_join.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/index_auto_self_join.q.out b/ql/src/test/results/beelinepositive/index_auto_self_join.q.out
deleted file mode 100644
index ec7e7f7..0000000
--- a/ql/src/test/results/beelinepositive/index_auto_self_join.q.out
+++ /dev/null
@@ -1,445 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/index_auto_self_join.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/index_auto_self_join.q
->>>  -- try the query without indexing, with manual indexing, and with automatic indexing
->>>  
->>>  EXPLAIN SELECT a.key, b.key FROM src a JOIN src b ON (a.value = b.value) WHERE a.key > 80 AND a.key < 100 AND b.key > 70 AND b.key < 90 ORDER BY a.key;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_TABREF (TOK_TABNAME src) a) (TOK_TABREF (TOK_TABNAME src) b) (= (. (TOK_TABLE_OR_COL a) value) (. (TOK_TABLE_OR_COL b) value)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL a) key)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL b) key))) (TOK_WHERE (AND (AND (AND (> (. (TOK_TABLE_OR_COL a) key) 80) (< (. (TOK_TABLE_OR_COL a) key) 100)) (> (. (TOK_TABLE_OR_COL b) key) 70)) (< (. (TOK_TABLE_OR_COL b) key) 90))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (. (TOK_TABLE_OR_COL a) key)))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-1 is a root stage'
-'  Stage-2 depends on stages: Stage-1'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        a '
-'          TableScan'
-'            alias: a'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key > 80.0) and (key < 100.0))'
-'                  type: boolean'
-'              Reduce Output Operator'
-'                key expressions:'
-'                      expr: value'
-'                      type: string'
-'                sort order: +'
-'                Map-reduce partition columns:'
-'                      expr: value'
-'                      type: string'
-'                tag: 0'
-'                value expressions:'
-'                      expr: key'
-'                      type: string'
-'        b '
-'          TableScan'
-'            alias: b'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key > 70.0) and (key < 90.0))'
-'                  type: boolean'
-'              Reduce Output Operator'
-'                key expressions:'
-'                      expr: value'
-'                      type: string'
-'                sort order: +'
-'                Map-reduce partition columns:'
-'                      expr: value'
-'                      type: string'
-'                tag: 1'
-'                value expressions:'
-'                      expr: key'
-'                      type: string'
-'      Reduce Operator Tree:'
-'        Join Operator'
-'          condition map:'
-'               Inner Join 0 to 1'
-'          condition expressions:'
-'            0 {VALUE._col0}'
-'            1 {VALUE._col0}'
-'          handleSkewJoin: false'
-'          outputColumnNames: _col0, _col4'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col0'
-'                  type: string'
-'                  expr: _col4'
-'                  type: string'
-'            outputColumnNames: _col0, _col1'
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-''
-'  Stage: Stage-2'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        file:!!{hive.exec.scratchdir}!! '
-'            Reduce Output Operator'
-'              key expressions:'
-'                    expr: _col0'
-'                    type: string'
-'              sort order: +'
-'              tag: -1'
-'              value expressions:'
-'                    expr: _col0'
-'                    type: string'
-'                    expr: _col1'
-'                    type: string'
-'      Reduce Operator Tree:'
-'        Extract'
-'          File Output Operator'
-'            compressed: false'
-'            GlobalTableId: 0'
-'            table:'
-'                input format: org.apache.hadoop.mapred.TextInputFormat'
-'                output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-''
-''
-102 rows selected 
->>>  SELECT a.key, b.key FROM src a JOIN src b ON (a.value = b.value) WHERE a.key > 80 AND a.key < 100 AND b.key > 70 AND b.key < 90 ORDER BY a.key;
-'key','key'
-'82','82'
-'83','83'
-'83','83'
-'83','83'
-'83','83'
-'84','84'
-'84','84'
-'84','84'
-'84','84'
-'85','85'
-'86','86'
-'87','87'
-12 rows selected 
->>>  
->>>  CREATE INDEX src_index ON TABLE src(key) as 'BITMAP' WITH DEFERRED REBUILD;
-No rows affected 
->>>  ALTER INDEX src_index ON src REBUILD;
-No rows affected 
->>>  
->>>  SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected 
->>>  SET hive.optimize.index.filter=true;
-No rows affected 
->>>  SET hive.optimize.index.filter.compact.minsize=0;
-No rows affected 
->>>  
->>>  EXPLAIN SELECT a.key, b.key FROM src a JOIN src b ON (a.value = b.value) WHERE a.key > 80 AND a.key < 100 AND b.key > 70 AND b.key < 90 ORDER BY a.key;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_TABREF (TOK_TABNAME src) a) (TOK_TABREF (TOK_TABNAME src) b) (= (. (TOK_TABLE_OR_COL a) value) (. (TOK_TABLE_OR_COL b) value)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL a) key)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL b) key))) (TOK_WHERE (AND (AND (AND (> (. (TOK_TABLE_OR_COL a) key) 80) (< (. (TOK_TABLE_OR_COL a) key) 100)) (> (. (TOK_TABLE_OR_COL b) key) 70)) (< (. (TOK_TABLE_OR_COL b) key) 90))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (. (TOK_TABLE_OR_COL a) key)))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-5 is a root stage'
-'  Stage-4 depends on stages: Stage-5'
-'  Stage-1 depends on stages: Stage-4, Stage-6'
-'  Stage-2 depends on stages: Stage-1'
-'  Stage-7 is a root stage'
-'  Stage-6 depends on stages: Stage-7'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-5'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        tmp_index:ind0:index_auto_self_join__src_src_index__ '
-'          TableScan'
-'            alias: index_auto_self_join__src_src_index__'
-'            filterExpr:'
-'                expr: (((key > 70.0) and (key < 90.0)) and (not EWAH_BITMAP_EMPTY(_bitmaps)))'
-'                type: boolean'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (((key > 70.0) and (key < 90.0)) and (not EWAH_BITMAP_EMPTY(_bitmaps)))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: _bucketname'
-'                      type: string'
-'                      expr: _offset'
-'                      type: bigint'
-'                      expr: _bitmaps'
-'                      type: array<bigint>'
-'                outputColumnNames: _col1, _col2, _col3'
-'                Select Operator'
-'                  expressions:'
-'                        expr: _col1'
-'                        type: string'
-'                        expr: _col2'
-'                        type: bigint'
-'                  outputColumnNames: _col0, _col1'
-'                  Select Operator'
-'                    expressions:'
-'                          expr: _col0'
-'                          type: string'
-'                          expr: _col1'
-'                          type: bigint'
-'                    outputColumnNames: _col0, _col1'
-'                    Group By Operator'
-'                      aggregations:'
-'                            expr: collect_set(_col1)'
-'                      bucketGroup: false'
-'                      keys:'
-'                            expr: _col0'
-'                            type: string'
-'                      mode: hash'
-'                      outputColumnNames: _col0, _col1'
-'                      Reduce Output Operator'
-'                        key expressions:'
-'                              expr: _col0'
-'                              type: string'
-'                        sort order: +'
-'                        Map-reduce partition columns:'
-'                              expr: _col0'
-'                              type: string'
-'                        tag: -1'
-'                        value expressions:'
-'                              expr: _col1'
-'                              type: array<bigint>'
-'      Reduce Operator Tree:'
-'        Group By Operator'
-'          aggregations:'
-'                expr: collect_set(VALUE._col0)'
-'          bucketGroup: false'
-'          keys:'
-'                expr: KEY._col0'
-'                type: string'
-'          mode: mergepartial'
-'          outputColumnNames: _col0, _col1'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col0'
-'                  type: string'
-'                  expr: _col1'
-'                  type: array<bigint>'
-'            outputColumnNames: _col0, _col1'
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 1'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-4'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          destination: file:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        a '
-'          TableScan'
-'            alias: a'
-'            filterExpr:'
-'                expr: ((key > 80.0) and (key < 100.0))'
-'                type: boolean'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key > 80.0) and (key < 100.0))'
-'                  type: boolean'
-'              Reduce Output Operator'
-'                key expressions:'
-'                      expr: value'
-'                      type: string'
-'                sort order: +'
-'                Map-reduce partition columns:'
-'                      expr: value'
-'                      type: string'
-'                tag: 0'
-'                value expressions:'
-'                      expr: key'
-'                      type: string'
-'        b '
-'          TableScan'
-'            alias: b'
-'            filterExpr:'
-'                expr: ((key > 70.0) and (key < 90.0))'
-'                type: boolean'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key > 70.0) and (key < 90.0))'
-'                  type: boolean'
-'              Reduce Output Operator'
-'                key expressions:'
-'                      expr: value'
-'                      type: string'
-'                sort order: +'
-'                Map-reduce partition columns:'
-'                      expr: value'
-'                      type: string'
-'                tag: 1'
-'                value expressions:'
-'                      expr: key'
-'                      type: string'
-'      Reduce Operator Tree:'
-'        Join Operator'
-'          condition map:'
-'               Inner Join 0 to 1'
-'          condition expressions:'
-'            0 {VALUE._col0}'
-'            1 {VALUE._col0}'
-'          handleSkewJoin: false'
-'          outputColumnNames: _col0, _col4'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col0'
-'                  type: string'
-'                  expr: _col4'
-'                  type: string'
-'            outputColumnNames: _col0, _col1'
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-''
-'  Stage: Stage-2'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        file:!!{hive.exec.scratchdir}!! '
-'            Reduce Output Operator'
-'              key expressions:'
-'                    expr: _col0'
-'                    type: string'
-'              sort order: +'
-'              tag: -1'
-'              value expressions:'
-'                    expr: _col0'
-'                    type: string'
-'                    expr: _col1'
-'                    type: string'
-'      Reduce Operator Tree:'
-'        Extract'
-'          File Output Operator'
-'            compressed: false'
-'            GlobalTableId: 0'
-'            table:'
-'                input format: org.apache.hadoop.mapred.TextInputFormat'
-'                output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-7'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        tmp_index:ind0:index_auto_self_join__src_src_index__ '
-'          TableScan'
-'            alias: index_auto_self_join__src_src_index__'
-'            filterExpr:'
-'                expr: (((key > 80.0) and (key < 100.0)) and (not EWAH_BITMAP_EMPTY(_bitmaps)))'
-'                type: boolean'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (((key > 80.0) and (key < 100.0)) and (not EWAH_BITMAP_EMPTY(_bitmaps)))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: _bucketname'
-'                      type: string'
-'                      expr: _offset'
-'                      type: bigint'
-'                      expr: _bitmaps'
-'                      type: array<bigint>'
-'                outputColumnNames: _col1, _col2, _col3'
-'                Select Operator'
-'                  expressions:'
-'                        expr: _col1'
-'                        type: string'
-'                        expr: _col2'
-'                        type: bigint'
-'                  outputColumnNames: _col0, _col1'
-'                  Select Operator'
-'                    expressions:'
-'                          expr: _col0'
-'                          type: string'
-'                          expr: _col1'
-'                          type: bigint'
-'                    outputColumnNames: _col0, _col1'
-'                    Group By Operator'
-'                      aggregations:'
-'                            expr: collect_set(_col1)'
-'                      bucketGroup: false'
-'                      keys:'
-'                            expr: _col0'
-'                            type: string'
-'                      mode: hash'
-'                      outputColumnNames: _col0, _col1'
-'                      Reduce Output Operator'
-'                        key expressions:'
-'                              expr: _col0'
-'                              type: string'
-'                        sort order: +'
-'                        Map-reduce partition columns:'
-'                              expr: _col0'
-'                              type: string'
-'                        tag: -1'
-'                        value expressions:'
-'                              expr: _col1'
-'                              type: array<bigint>'
-'      Reduce Operator Tree:'
-'        Group By Operator'
-'          aggregations:'
-'                expr: collect_set(VALUE._col0)'
-'          bucketGroup: false'
-'          keys:'
-'                expr: KEY._col0'
-'                type: string'
-'          mode: mergepartial'
-'          outputColumnNames: _col0, _col1'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col0'
-'                  type: string'
-'                  expr: _col1'
-'                  type: array<bigint>'
-'            outputColumnNames: _col0, _col1'
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 1'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-6'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          destination: file:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-''
-''
-286 rows selected 
->>>  SELECT a.key, b.key FROM src a JOIN src b ON (a.value = b.value) WHERE a.key > 80 AND a.key < 100 AND b.key > 70 AND b.key < 90 ORDER BY a.key;
-'key','key'
-'82','82'
-'83','83'
-'83','83'
-'83','83'
-'83','83'
-'84','84'
-'84','84'
-'84','84'
-'84','84'
-'85','85'
-'86','86'
-'87','87'
-12 rows selected 
->>>  
->>>  DROP INDEX src_index on src;
-No rows affected 
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/index_auto_unused.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/index_auto_unused.q.out b/ql/src/test/results/beelinepositive/index_auto_unused.q.out
deleted file mode 100644
index cc2a607..0000000
--- a/ql/src/test/results/beelinepositive/index_auto_unused.q.out
+++ /dev/null
@@ -1,484 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/index_auto_unused.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/index_auto_unused.q
->>>  -- test cases where the index should not be used automatically
->>>  
->>>  CREATE INDEX src_index ON TABLE src(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected 
->>>  ALTER INDEX src_index ON src REBUILD;
-No rows affected 
->>>  
->>>  SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected 
->>>  SET hive.optimize.index.filter=true;
-No rows affected 
->>>  SET hive.optimize.index.filter.compact.minsize=5368709120;
-No rows affected 
->>>  SET hive.optimize.index.filter.compact.maxsize=-1;
-No rows affected 
->>>  
->>>  -- min size too large (src is less than 5G)
->>>  EXPLAIN SELECT * FROM src WHERE key > 80 AND key < 100 ORDER BY key;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (AND (> (TOK_TABLE_OR_COL key) 80) (< (TOK_TABLE_OR_COL key) 100))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-1 is a root stage'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src '
-'          TableScan'
-'            alias: src'
-'            filterExpr:'
-'                expr: ((key > 80.0) and (key < 100.0))'
-'                type: boolean'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key > 80.0) and (key < 100.0))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                Reduce Output Operator'
-'                  key expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                  sort order: +'
-'                  tag: -1'
-'                  value expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                        expr: _col1'
-'                        type: string'
-'      Reduce Operator Tree:'
-'        Extract'
-'          File Output Operator'
-'            compressed: false'
-'            GlobalTableId: 0'
-'            table:'
-'                input format: org.apache.hadoop.mapred.TextInputFormat'
-'                output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-''
-''
-53 rows selected 
->>>  SELECT * FROM src WHERE key > 80 AND key < 100 ORDER BY key;
-'key','value'
-'82','val_82'
-'83','val_83'
-'83','val_83'
-'84','val_84'
-'84','val_84'
-'85','val_85'
-'86','val_86'
-'87','val_87'
-'90','val_90'
-'90','val_90'
-'90','val_90'
-'92','val_92'
-'95','val_95'
-'95','val_95'
-'96','val_96'
-'97','val_97'
-'97','val_97'
-'98','val_98'
-'98','val_98'
-19 rows selected 
->>>  
->>>  SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected 
->>>  SET hive.optimize.index.filter=true;
-No rows affected 
->>>  SET hive.optimize.index.filter.compact.minsize=0;
-No rows affected 
->>>  SET hive.optimize.index.filter.compact.maxsize=1;
-No rows affected 
->>>  
->>>  -- max size too small
->>>  EXPLAIN SELECT * FROM src WHERE key > 80 AND key < 100 ORDER BY key;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (AND (> (TOK_TABLE_OR_COL key) 80) (< (TOK_TABLE_OR_COL key) 100))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-1 is a root stage'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src '
-'          TableScan'
-'            alias: src'
-'            filterExpr:'
-'                expr: ((key > 80.0) and (key < 100.0))'
-'                type: boolean'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key > 80.0) and (key < 100.0))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                Reduce Output Operator'
-'                  key expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                  sort order: +'
-'                  tag: -1'
-'                  value expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                        expr: _col1'
-'                        type: string'
-'      Reduce Operator Tree:'
-'        Extract'
-'          File Output Operator'
-'            compressed: false'
-'            GlobalTableId: 0'
-'            table:'
-'                input format: org.apache.hadoop.mapred.TextInputFormat'
-'                output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-''
-''
-53 rows selected 
->>>  SELECT * FROM src WHERE key > 80 AND key < 100 ORDER BY key;
-'key','value'
-'82','val_82'
-'83','val_83'
-'83','val_83'
-'84','val_84'
-'84','val_84'
-'85','val_85'
-'86','val_86'
-'87','val_87'
-'90','val_90'
-'90','val_90'
-'90','val_90'
-'92','val_92'
-'95','val_95'
-'95','val_95'
-'96','val_96'
-'97','val_97'
-'97','val_97'
-'98','val_98'
-'98','val_98'
-19 rows selected 
->>>  
->>>  SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected 
->>>  SET hive.optimize.index.filter=true;
-No rows affected 
->>>  SET hive.optimize.index.filter.compact.minsize=0;
-No rows affected 
->>>  SET hive.optimize.index.filter.compact.maxsize=-1;
-No rows affected 
->>>  
->>>  -- OR predicate not supported by compact indexes
->>>  EXPLAIN SELECT * FROM src WHERE key < 10 OR key > 480 ORDER BY key;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (OR (< (TOK_TABLE_OR_COL key) 10) (> (TOK_TABLE_OR_COL key) 480))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-1 is a root stage'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src '
-'          TableScan'
-'            alias: src'
-'            filterExpr:'
-'                expr: ((key < 10.0) or (key > 480.0))'
-'                type: boolean'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key < 10.0) or (key > 480.0))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                Reduce Output Operator'
-'                  key expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                  sort order: +'
-'                  tag: -1'
-'                  value expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                        expr: _col1'
-'                        type: string'
-'      Reduce Operator Tree:'
-'        Extract'
-'          File Output Operator'
-'            compressed: false'
-'            GlobalTableId: 0'
-'            table:'
-'                input format: org.apache.hadoop.mapred.TextInputFormat'
-'                output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-''
-''
-53 rows selected 
->>>  SELECT * FROM src WHERE key < 10 OR key > 480 ORDER BY key;
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-'2','val_2'
-'4','val_4'
-'481','val_481'
-'482','val_482'
-'483','val_483'
-'484','val_484'
-'485','val_485'
-'487','val_487'
-'489','val_489'
-'489','val_489'
-'489','val_489'
-'489','val_489'
-'490','val_490'
-'491','val_491'
-'492','val_492'
-'492','val_492'
-'493','val_493'
-'494','val_494'
-'495','val_495'
-'496','val_496'
-'497','val_497'
-'498','val_498'
-'498','val_498'
-'498','val_498'
-'5','val_5'
-'5','val_5'
-'5','val_5'
-'8','val_8'
-'9','val_9'
-32 rows selected 
->>>  
->>>  SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected 
->>>  SET hive.optimize.index.filter=true;
-No rows affected 
->>>  SET hive.optimize.index.filter.compact.minsize=0;
-No rows affected 
->>>  SET hive.optimize.index.filter.compact.maxsize=-1;
-No rows affected 
->>>  
->>>  -- columns are not covered by indexes
->>>  DROP INDEX src_index on src;
-No rows affected 
->>>  CREATE INDEX src_val_index ON TABLE src(value) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected 
->>>  ALTER INDEX src_val_index ON src REBUILD;
-No rows affected 
->>>  
->>>  EXPLAIN SELECT * FROM src WHERE key > 80 AND key < 100 ORDER BY key;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (AND (> (TOK_TABLE_OR_COL key) 80) (< (TOK_TABLE_OR_COL key) 100))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-1 is a root stage'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src '
-'          TableScan'
-'            alias: src'
-'            filterExpr:'
-'                expr: ((key > 80.0) and (key < 100.0))'
-'                type: boolean'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key > 80.0) and (key < 100.0))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                Reduce Output Operator'
-'                  key expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                  sort order: +'
-'                  tag: -1'
-'                  value expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                        expr: _col1'
-'                        type: string'
-'      Reduce Operator Tree:'
-'        Extract'
-'          File Output Operator'
-'            compressed: false'
-'            GlobalTableId: 0'
-'            table:'
-'                input format: org.apache.hadoop.mapred.TextInputFormat'
-'                output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-''
-''
-53 rows selected 
->>>  SELECT * FROM src WHERE key > 80 AND key < 100 ORDER BY key;
-'key','value'
-'82','val_82'
-'83','val_83'
-'83','val_83'
-'84','val_84'
-'84','val_84'
-'85','val_85'
-'86','val_86'
-'87','val_87'
-'90','val_90'
-'90','val_90'
-'90','val_90'
-'92','val_92'
-'95','val_95'
-'95','val_95'
-'96','val_96'
-'97','val_97'
-'97','val_97'
-'98','val_98'
-'98','val_98'
-19 rows selected 
->>>  
->>>  DROP INDEX src_val_index on src;
-No rows affected 
->>>  
->>>  SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected 
->>>  SET hive.optimize.index.filter=true;
-No rows affected 
->>>  SET hive.optimize.index.filter.compact.minsize=0;
-No rows affected 
->>>  SET hive.optimize.index.filter.compact.maxsize=-1;
-No rows affected 
->>>  
->>>  -- required partitions have not been built yet
->>>  CREATE INDEX src_part_index ON TABLE srcpart(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected 
->>>  ALTER INDEX src_part_index ON srcpart PARTITION (ds='2008-04-08', hr=11) REBUILD;
-No rows affected 
->>>  
->>>  EXPLAIN SELECT * FROM srcpart WHERE ds='2008-04-09' AND hr=12 AND key < 10 ORDER BY key;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME srcpart))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (AND (AND (= (TOK_TABLE_OR_COL ds) '2008-04-09') (= (TOK_TABLE_OR_COL hr) 12)) (< (TOK_TABLE_OR_COL key) 10))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-1 is a root stage'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        srcpart '
-'          TableScan'
-'            alias: srcpart'
-'            filterExpr:'
-'                expr: (((ds = '2008-04-09') and (hr = 12.0)) and (key < 10.0))'
-'                type: boolean'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key < 10.0)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                      expr: ds'
-'                      type: string'
-'                      expr: hr'
-'                      type: string'
-'                outputColumnNames: _col0, _col1, _col2, _col3'
-'                Reduce Output Operator'
-'                  key expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                  sort order: +'
-'                  tag: -1'
-'                  value expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                        expr: _col1'
-'                        type: string'
-'                        expr: _col2'
-'                        type: string'
-'                        expr: _col3'
-'                        type: string'
-'      Reduce Operator Tree:'
-'        Extract'
-'          File Output Operator'
-'            compressed: false'
-'            GlobalTableId: 0'
-'            table:'
-'                input format: org.apache.hadoop.mapred.TextInputFormat'
-'                output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-''
-''
-61 rows selected 
->>>  SELECT * FROM srcpart WHERE ds='2008-04-09' AND hr=12 AND key < 10 ORDER BY key;
-'key','value','ds','hr'
-'0','val_0','2008-04-09','12'
-'0','val_0','2008-04-09','12'
-'0','val_0','2008-04-09','12'
-'2','val_2','2008-04-09','12'
-'4','val_4','2008-04-09','12'
-'5','val_5','2008-04-09','12'
-'5','val_5','2008-04-09','12'
-'5','val_5','2008-04-09','12'
-'8','val_8','2008-04-09','12'
-'9','val_9','2008-04-09','12'
-10 rows selected 
->>>  
->>>  DROP INDEX src_part_index on srcpart;
-No rows affected 
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/index_auto_update.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/index_auto_update.q.out b/ql/src/test/results/beelinepositive/index_auto_update.q.out
deleted file mode 100644
index 3330998..0000000
--- a/ql/src/test/results/beelinepositive/index_auto_update.q.out
+++ /dev/null
@@ -1,352 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/index_auto_update.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/index_auto_update.q
->>>  -- Test if index is actually being used.
->>>  
->>>  -- Create temp, and populate it with some values in src.
->>>  CREATE TABLE temp(key STRING, val STRING) STORED AS TEXTFILE;
-No rows affected 
->>>  INSERT OVERWRITE TABLE temp SELECT * FROM src WHERE key < 50;
-'key','value'
-No rows selected 
->>>  
->>>  -- Build an index on temp.
->>>  CREATE INDEX temp_index ON TABLE temp(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected 
->>>  ALTER INDEX temp_index ON temp REBUILD;
-No rows affected 
->>>  
->>>  SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected 
->>>  SET hive.optimize.index.filter=true;
-No rows affected 
->>>  SET hive.optimize.index.autoupdate=true;
-No rows affected 
->>>  SET hive.optimize.index.filter.compact.minsize=0;
-No rows affected 
->>>  
->>>  -- overwrite temp table so index is out of date
->>>  EXPLAIN INSERT OVERWRITE TABLE temp SELECT * FROM src;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME temp))) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-4 is a root stage'
-'  Stage-10 depends on stages: Stage-4 , consists of Stage-7, Stage-6, Stage-8'
-'  Stage-7'
-'  Stage-0 depends on stages: Stage-7, Stage-6, Stage-9'
-'  Stage-1 depends on stages: Stage-0'
-'  Stage-0 depends on stages: Stage-1'
-'  Stage-2 depends on stages: Stage-0'
-'  null depends on stages: Stage-1'
-'  Stage-3 depends on stages: Stage-1'
-'  Stage-5 depends on stages: Stage-0'
-'  Stage-6'
-'  Stage-8'
-'  Stage-9 depends on stages: Stage-8'
-''
-'STAGE PLANS:'
-'  Stage: Stage-4'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src '
-'          TableScan'
-'            alias: src'
-'            Select Operator'
-'              expressions:'
-'                    expr: key'
-'                    type: string'
-'                    expr: value'
-'                    type: string'
-'              outputColumnNames: _col0, _col1'
-'              File Output Operator'
-'                compressed: false'
-'                GlobalTableId: 1'
-'                table:'
-'                    input format: org.apache.hadoop.mapred.TextInputFormat'
-'                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                    name: index_auto_update.temp'
-''
-'  Stage: Stage-10'
-'    Conditional Operator'
-''
-'  Stage: Stage-7'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          destination: pfile:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-0'
-'    Move Operator'
-'      tables:'
-'          replace: true'
-'          table:'
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: index_auto_update.temp'
-''
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        temp '
-'          TableScan'
-'            alias: temp'
-'            Select Operator'
-'              expressions:'
-'                    expr: key'
-'                    type: string'
-'                    expr: INPUT__FILE__NAME'
-'                    type: string'
-'                    expr: BLOCK__OFFSET__INSIDE__FILE'
-'                    type: bigint'
-'              outputColumnNames: key, INPUT__FILE__NAME, BLOCK__OFFSET__INSIDE__FILE'
-'              Group By Operator'
-'                aggregations:'
-'                      expr: collect_set(BLOCK__OFFSET__INSIDE__FILE)'
-'                bucketGroup: false'
-'                keys:'
-'                      expr: key'
-'                      type: string'
-'                      expr: INPUT__FILE__NAME'
-'                      type: string'
-'                mode: hash'
-'                outputColumnNames: _col0, _col1, _col2'
-'                Reduce Output Operator'
-'                  key expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                        expr: _col1'
-'                        type: string'
-'                  sort order: ++'
-'                  Map-reduce partition columns:'
-'                        expr: _col0'
-'                        type: string'
-'                        expr: _col1'
-'                        type: string'
-'                  tag: -1'
-'                  value expressions:'
-'                        expr: _col2'
-'                        type: array<bigint>'
-'      Reduce Operator Tree:'
-'        Group By Operator'
-'          aggregations:'
-'                expr: collect_set(VALUE._col0)'
-'          bucketGroup: false'
-'          keys:'
-'                expr: KEY._col0'
-'                type: string'
-'                expr: KEY._col1'
-'                type: string'
-'          mode: mergepartial'
-'          outputColumnNames: _col0, _col1, _col2'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col0'
-'                  type: string'
-'                  expr: _col1'
-'                  type: string'
-'                  expr: _col2'
-'                  type: array<bigint>'
-'            outputColumnNames: _col0, _col1, _col2'
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 1'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                  name: index_auto_update.index_auto_update__temp_temp_index__'
-''
-'  Stage: Stage-0'
-'    Move Operator'
-'      tables:'
-'          replace: true'
-'          table:'
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: index_auto_update.index_auto_update__temp_temp_index__'
-''
-'  Stage: Stage-2'
-'    Stats-Aggr Operator'
-''
-'  Stage: null'
-''
-'  Stage: Stage-3'
-''
-'  Stage: Stage-5'
-'    Stats-Aggr Operator'
-''
-'  Stage: Stage-6'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        pfile:!!{hive.exec.scratchdir}!! '
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                  name: index_auto_update.temp'
-''
-'  Stage: Stage-8'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        pfile:!!{hive.exec.scratchdir}!! '
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                  name: index_auto_update.temp'
-''
-'  Stage: Stage-9'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          destination: pfile:!!{hive.exec.scratchdir}!!'
-''
-''
-185 rows selected 
->>>  INSERT OVERWRITE TABLE temp SELECT * FROM src;
-'key','value'
-No rows selected 
->>>  
->>>  -- query should return indexed values
->>>  EXPLAIN SELECT * FROM temp WHERE key  = 86;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME temp))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (= (TOK_TABLE_OR_COL key) 86))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-3 is a root stage'
-'  Stage-8 depends on stages: Stage-3 , consists of Stage-5, Stage-4, Stage-6'
-'  Stage-5'
-'  Stage-2 depends on stages: Stage-5, Stage-4, Stage-7'
-'  Stage-1 depends on stages: Stage-2'
-'  Stage-4'
-'  Stage-6'
-'  Stage-7 depends on stages: Stage-6'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-3'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        index_auto_update__temp_temp_index__ '
-'          TableScan'
-'            alias: index_auto_update__temp_temp_index__'
-'            filterExpr:'
-'                expr: (key = 86.0)'
-'                type: boolean'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key = 86.0)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: _bucketname'
-'                      type: string'
-'                      expr: _offsets'
-'                      type: array<bigint>'
-'                outputColumnNames: _col0, _col1'
-'                File Output Operator'
-'                  compressed: false'
-'                  GlobalTableId: 1'
-'                  table:'
-'                      input format: org.apache.hadoop.mapred.TextInputFormat'
-'                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-8'
-'    Conditional Operator'
-''
-'  Stage: Stage-5'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          destination: file:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-2'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          destination: file:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        temp '
-'          TableScan'
-'            alias: temp'
-'            filterExpr:'
-'                expr: (key = 86.0)'
-'                type: boolean'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key = 86.0)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: val'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                File Output Operator'
-'                  compressed: false'
-'                  GlobalTableId: 0'
-'                  table:'
-'                      input format: org.apache.hadoop.mapred.TextInputFormat'
-'                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-4'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        file:!!{hive.exec.scratchdir}!! '
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-6'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        file:!!{hive.exec.scratchdir}!! '
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-7'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          destination: file:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-''
-''
-117 rows selected 
->>>  SELECT * FROM temp WHERE key  = 86;
-'key','val'
-'86','val_86'
-1 row selected 
->>>  
->>>  SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected 
->>>  SET hive.optimize.index.filter=false;
-No rows affected 
->>>  DROP table temp;
-No rows affected 
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/index_bitmap_auto_partitioned.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/index_bitmap_auto_partitioned.q.out b/ql/src/test/results/beelinepositive/index_bitmap_auto_partitioned.q.out
deleted file mode 100644
index 3dcd175..0000000
--- a/ql/src/test/results/beelinepositive/index_bitmap_auto_partitioned.q.out
+++ /dev/null
@@ -1,167 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/index_bitmap_auto_partitioned.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/index_bitmap_auto_partitioned.q
->>>  -- test automatic use of index on table with partitions
->>>  CREATE INDEX src_part_index ON TABLE srcpart(key) as 'BITMAP' WITH DEFERRED REBUILD;
-No rows affected 
->>>  ALTER INDEX src_part_index ON srcpart REBUILD;
-No rows affected 
->>>  
->>>  SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected 
->>>  SET hive.optimize.index.filter=true;
-No rows affected 
->>>  
->>>  EXPLAIN SELECT key, value FROM srcpart WHERE key=86 AND ds='2008-04-09' ORDER BY key;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME srcpart))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_TABLE_OR_COL value))) (TOK_WHERE (AND (= (TOK_TABLE_OR_COL key) 86) (= (TOK_TABLE_OR_COL ds) '2008-04-09'))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-3 is a root stage'
-'  Stage-2 depends on stages: Stage-3'
-'  Stage-1 depends on stages: Stage-2'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-3'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        tmp_index:ind0:index_bitmap_auto_partitioned__srcpart_src_part_index__ '
-'          TableScan'
-'            alias: index_bitmap_auto_partitioned__srcpart_src_part_index__'
-'            filterExpr:'
-'                expr: ((key = 86.0) and (not EWAH_BITMAP_EMPTY(_bitmaps)))'
-'                type: boolean'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key = 86.0) and (not EWAH_BITMAP_EMPTY(_bitmaps)))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: _bucketname'
-'                      type: string'
-'                      expr: _offset'
-'                      type: bigint'
-'                      expr: _bitmaps'
-'                      type: array<bigint>'
-'                outputColumnNames: _col1, _col2, _col3'
-'                Select Operator'
-'                  expressions:'
-'                        expr: _col1'
-'                        type: string'
-'                        expr: _col2'
-'                        type: bigint'
-'                  outputColumnNames: _col0, _col1'
-'                  Select Operator'
-'                    expressions:'
-'                          expr: _col0'
-'                          type: string'
-'                          expr: _col1'
-'                          type: bigint'
-'                    outputColumnNames: _col0, _col1'
-'                    Group By Operator'
-'                      aggregations:'
-'                            expr: collect_set(_col1)'
-'                      bucketGroup: false'
-'                      keys:'
-'                            expr: _col0'
-'                            type: string'
-'                      mode: hash'
-'                      outputColumnNames: _col0, _col1'
-'                      Reduce Output Operator'
-'                        key expressions:'
-'                              expr: _col0'
-'                              type: string'
-'                        sort order: +'
-'                        Map-reduce partition columns:'
-'                              expr: _col0'
-'                              type: string'
-'                        tag: -1'
-'                        value expressions:'
-'                              expr: _col1'
-'                              type: array<bigint>'
-'      Reduce Operator Tree:'
-'        Group By Operator'
-'          aggregations:'
-'                expr: collect_set(VALUE._col0)'
-'          bucketGroup: false'
-'          keys:'
-'                expr: KEY._col0'
-'                type: string'
-'          mode: mergepartial'
-'          outputColumnNames: _col0, _col1'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col0'
-'                  type: string'
-'                  expr: _col1'
-'                  type: array<bigint>'
-'            outputColumnNames: _col0, _col1'
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 1'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-2'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          destination: file:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        srcpart '
-'          TableScan'
-'            alias: srcpart'
-'            filterExpr:'
-'                expr: ((key = 86.0) and (ds = '2008-04-09'))'
-'                type: boolean'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key = 86.0)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                Reduce Output Operator'
-'                  key expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                  sort order: +'
-'                  tag: -1'
-'                  value expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                        expr: _col1'
-'                        type: string'
-'      Reduce Operator Tree:'
-'        Extract'
-'          File Output Operator'
-'            compressed: false'
-'            GlobalTableId: 0'
-'            table:'
-'                input format: org.apache.hadoop.mapred.TextInputFormat'
-'                output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-''
-''
-142 rows selected 
->>>  SELECT key, value FROM srcpart WHERE key=86 AND ds='2008-04-09' ORDER BY key;
-'key','value'
-'86','val_86'
-'86','val_86'
-2 rows selected 
->>>  
->>>  DROP INDEX src_part_index ON srcpart;
-No rows affected 
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/index_bitmap_compression.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/index_bitmap_compression.q.out b/ql/src/test/results/beelinepositive/index_bitmap_compression.q.out
deleted file mode 100644
index 9bf0106..0000000
--- a/ql/src/test/results/beelinepositive/index_bitmap_compression.q.out
+++ /dev/null
@@ -1,188 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/index_bitmap_compression.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/index_bitmap_compression.q
->>>  SET hive.exec.compress.result=true;
-No rows affected 
->>>  CREATE INDEX src_index ON TABLE src(key) as 'BITMAP' WITH DEFERRED REBUILD;
-No rows affected 
->>>  ALTER INDEX src_index ON src REBUILD;
-No rows affected 
->>>  
->>>  SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected 
->>>  SET hive.optimize.index.filter=true;
-No rows affected 
->>>  SET hive.optimize.index.filter.compact.minsize=0;
-No rows affected 
->>>  
->>>  -- automatic indexing
->>>  EXPLAIN SELECT key, value FROM src WHERE key > 80 AND key < 100 ORDER BY key;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_TABLE_OR_COL value))) (TOK_WHERE (AND (> (TOK_TABLE_OR_COL key) 80) (< (TOK_TABLE_OR_COL key) 100))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-3 is a root stage'
-'  Stage-2 depends on stages: Stage-3'
-'  Stage-1 depends on stages: Stage-2'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-3'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        tmp_index:ind0:index_bitmap_compression__src_src_index__ '
-'          TableScan'
-'            alias: index_bitmap_compression__src_src_index__'
-'            filterExpr:'
-'                expr: (((key > 80.0) and (key < 100.0)) and (not EWAH_BITMAP_EMPTY(_bitmaps)))'
-'                type: boolean'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (((key > 80.0) and (key < 100.0)) and (not EWAH_BITMAP_EMPTY(_bitmaps)))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: _bucketname'
-'                      type: string'
-'                      expr: _offset'
-'                      type: bigint'
-'                      expr: _bitmaps'
-'                      type: array<bigint>'
-'                outputColumnNames: _col1, _col2, _col3'
-'                Select Operator'
-'                  expressions:'
-'                        expr: _col1'
-'                        type: string'
-'                        expr: _col2'
-'                        type: bigint'
-'                  outputColumnNames: _col0, _col1'
-'                  Select Operator'
-'                    expressions:'
-'                          expr: _col0'
-'                          type: string'
-'                          expr: _col1'
-'                          type: bigint'
-'                    outputColumnNames: _col0, _col1'
-'                    Group By Operator'
-'                      aggregations:'
-'                            expr: collect_set(_col1)'
-'                      bucketGroup: false'
-'                      keys:'
-'                            expr: _col0'
-'                            type: string'
-'                      mode: hash'
-'                      outputColumnNames: _col0, _col1'
-'                      Reduce Output Operator'
-'                        key expressions:'
-'                              expr: _col0'
-'                              type: string'
-'                        sort order: +'
-'                        Map-reduce partition columns:'
-'                              expr: _col0'
-'                              type: string'
-'                        tag: -1'
-'                        value expressions:'
-'                              expr: _col1'
-'                              type: array<bigint>'
-'      Reduce Operator Tree:'
-'        Group By Operator'
-'          aggregations:'
-'                expr: collect_set(VALUE._col0)'
-'          bucketGroup: false'
-'          keys:'
-'                expr: KEY._col0'
-'                type: string'
-'          mode: mergepartial'
-'          outputColumnNames: _col0, _col1'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col0'
-'                  type: string'
-'                  expr: _col1'
-'                  type: array<bigint>'
-'            outputColumnNames: _col0, _col1'
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 1'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-2'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          destination: file:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src '
-'          TableScan'
-'            alias: src'
-'            filterExpr:'
-'                expr: ((key > 80.0) and (key < 100.0))'
-'                type: boolean'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key > 80.0) and (key < 100.0))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                Reduce Output Operator'
-'                  key expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                  sort order: +'
-'                  tag: -1'
-'                  value expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                        expr: _col1'
-'                        type: string'
-'      Reduce Operator Tree:'
-'        Extract'
-'          File Output Operator'
-'            compressed: false'
-'            GlobalTableId: 0'
-'            table:'
-'                input format: org.apache.hadoop.mapred.TextInputFormat'
-'                output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-''
-''
-142 rows selected 
->>>  SELECT key, value FROM src WHERE key > 80 AND key < 100 ORDER BY key;
-'key','value'
-'82','val_82'
-'83','val_83'
-'83','val_83'
-'84','val_84'
-'84','val_84'
-'85','val_85'
-'86','val_86'
-'87','val_87'
-'90','val_90'
-'90','val_90'
-'90','val_90'
-'92','val_92'
-'95','val_95'
-'95','val_95'
-'96','val_96'
-'97','val_97'
-'97','val_97'
-'98','val_98'
-'98','val_98'
-19 rows selected 
->>>  
->>>  DROP INDEX src_index on src;
-No rows affected 
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/index_compact_binary_search.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/index_compact_binary_search.q.out b/ql/src/test/results/beelinepositive/index_compact_binary_search.q.out
deleted file mode 100644
index e963587..0000000
--- a/ql/src/test/results/beelinepositive/index_compact_binary_search.q.out
+++ /dev/null
@@ -1,489 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/index_compact_binary_search.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/index_compact_binary_search.q
->>>  SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected 
->>>  SET hive.default.fileformat=TextFile;
-No rows affected 
->>>  
->>>  CREATE INDEX src_index ON TABLE src(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected 
->>>  ALTER INDEX src_index ON src REBUILD;
-No rows affected 
->>>  
->>>  SET hive.optimize.index.filter=true;
-No rows affected 
->>>  SET hive.optimize.index.filter.compact.minsize=1;
-No rows affected 
->>>  SET hive.index.compact.binary.search=true;
-No rows affected 
->>>  
->>>  SET hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.VerifyHiveSortedInputFormatUsedHook;
-No rows affected 
->>>  
->>>  SELECT * FROM src WHERE key = '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key < '1';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key <= '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key > '8';
-'key','value'
-'86','val_86'
-'98','val_98'
-'82','val_82'
-'92','val_92'
-'83','val_83'
-'84','val_84'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'85','val_85'
-'87','val_87'
-'90','val_90'
-'95','val_95'
-'80','val_80'
-'90','val_90'
-'83','val_83'
-'9','val_9'
-'97','val_97'
-'84','val_84'
-'90','val_90'
-'97','val_97'
-21 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key >= '9';
-'key','value'
-'98','val_98'
-'92','val_92'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'90','val_90'
-'95','val_95'
-'90','val_90'
-'9','val_9'
-'97','val_97'
-'90','val_90'
-'97','val_97'
-12 rows selected 
->>>  
->>>  SET hive.exec.post.hooks=;
-No rows affected 
->>>  
->>>  DROP INDEX src_index ON src;
-No rows affected 
->>>  
->>>  SET hive.default.fileformat=RCFILE;
-No rows affected 
->>>  
->>>  CREATE INDEX src_index ON TABLE src(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected 
->>>  ALTER INDEX src_index ON src REBUILD;
-No rows affected 
->>>  
->>>  SET hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.VerifyHiveSortedInputFormatUsedHook;
-No rows affected 
->>>  
->>>  SELECT * FROM src WHERE key = '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key < '1';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key <= '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key > '8';
-'key','value'
-'86','val_86'
-'98','val_98'
-'82','val_82'
-'92','val_92'
-'83','val_83'
-'84','val_84'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'85','val_85'
-'87','val_87'
-'90','val_90'
-'95','val_95'
-'80','val_80'
-'90','val_90'
-'83','val_83'
-'9','val_9'
-'97','val_97'
-'84','val_84'
-'90','val_90'
-'97','val_97'
-21 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key >= '9';
-'key','value'
-'98','val_98'
-'92','val_92'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'90','val_90'
-'95','val_95'
-'90','val_90'
-'9','val_9'
-'97','val_97'
-'90','val_90'
-'97','val_97'
-12 rows selected 
->>>  
->>>  SET hive.exec.post.hooks=;
-No rows affected 
->>>  
->>>  DROP INDEX src_index ON src;
-No rows affected 
->>>  
->>>  SET hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat;
-No rows affected 
->>>  SET hive.default.fileformat=TextFile;
-No rows affected 
->>>  
->>>  CREATE INDEX src_index ON TABLE src(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected 
->>>  ALTER INDEX src_index ON src REBUILD;
-No rows affected 
->>>  
->>>  SET hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.VerifyHiveSortedInputFormatUsedHook;
-No rows affected 
->>>  
->>>  SELECT * FROM src WHERE key = '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key < '1';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key <= '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key > '8';
-'key','value'
-'86','val_86'
-'98','val_98'
-'82','val_82'
-'92','val_92'
-'83','val_83'
-'84','val_84'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'85','val_85'
-'87','val_87'
-'90','val_90'
-'95','val_95'
-'80','val_80'
-'90','val_90'
-'83','val_83'
-'9','val_9'
-'97','val_97'
-'84','val_84'
-'90','val_90'
-'97','val_97'
-21 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key >= '9';
-'key','value'
-'98','val_98'
-'92','val_92'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'90','val_90'
-'95','val_95'
-'90','val_90'
-'9','val_9'
-'97','val_97'
-'90','val_90'
-'97','val_97'
-12 rows selected 
->>>  
->>>  SET hive.exec.post.hooks=;
-No rows affected 
->>>  
->>>  DROP INDEX src_index ON src;
-No rows affected 
->>>  
->>>  SET hive.default.fileformat=RCFILE;
-No rows affected 
->>>  
->>>  CREATE INDEX src_index ON TABLE src(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected 
->>>  ALTER INDEX src_index ON src REBUILD;
-No rows affected 
->>>  
->>>  SET hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.VerifyHiveSortedInputFormatUsedHook;
-No rows affected 
->>>  
->>>  SELECT * FROM src WHERE key = '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key < '1';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key <= '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key > '8';
-'key','value'
-'86','val_86'
-'98','val_98'
-'82','val_82'
-'92','val_92'
-'83','val_83'
-'84','val_84'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'85','val_85'
-'87','val_87'
-'90','val_90'
-'95','val_95'
-'80','val_80'
-'90','val_90'
-'83','val_83'
-'9','val_9'
-'97','val_97'
-'84','val_84'
-'90','val_90'
-'97','val_97'
-21 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key >= '9';
-'key','value'
-'98','val_98'
-'92','val_92'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'90','val_90'
-'95','val_95'
-'90','val_90'
-'9','val_9'
-'97','val_97'
-'90','val_90'
-'97','val_97'
-12 rows selected 
->>>  
->>>  SET hive.exec.post.hooks=;
-No rows affected 
->>>  
->>>  DROP INDEX src_index ON src;
-No rows affected 
->>>  
->>>  SET hive.input.format=org.apache.hadoop.hive.ql.io.BucketizedHiveInputFormat;
-No rows affected 
->>>  SET hive.default.fileformat=TextFile;
-No rows affected 
->>>  
->>>  CREATE INDEX src_index ON TABLE src(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected 
->>>  ALTER INDEX src_index ON src REBUILD;
-No rows affected 
->>>  
->>>  SET hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.VerifyHiveSortedInputFormatUsedHook;
-No rows affected 
->>>  
->>>  SELECT * FROM src WHERE key = '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key < '1';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key <= '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key > '8';
-'key','value'
-'86','val_86'
-'98','val_98'
-'82','val_82'
-'92','val_92'
-'83','val_83'
-'84','val_84'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'85','val_85'
-'87','val_87'
-'90','val_90'
-'95','val_95'
-'80','val_80'
-'90','val_90'
-'83','val_83'
-'9','val_9'
-'97','val_97'
-'84','val_84'
-'90','val_90'
-'97','val_97'
-21 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key >= '9';
-'key','value'
-'98','val_98'
-'92','val_92'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'90','val_90'
-'95','val_95'
-'90','val_90'
-'9','val_9'
-'97','val_97'
-'90','val_90'
-'97','val_97'
-12 rows selected 
->>>  
->>>  SET hive.exec.post.hooks=;
-No rows affected 
->>>  
->>>  DROP INDEX src_index ON src;
-No rows affected 
->>>  
->>>  SET hive.default.fileformat=RCFILE;
-No rows affected 
->>>  
->>>  CREATE INDEX src_index ON TABLE src(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected 
->>>  ALTER INDEX src_index ON src REBUILD;
-No rows affected 
->>>  
->>>  SET hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.VerifyHiveSortedInputFormatUsedHook;
-No rows affected 
->>>  
->>>  SELECT * FROM src WHERE key = '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key < '1';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key <= '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key > '8';
-'key','value'
-'86','val_86'
-'98','val_98'
-'82','val_82'
-'92','val_92'
-'83','val_83'
-'84','val_84'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'85','val_85'
-'87','val_87'
-'90','val_90'
-'95','val_95'
-'80','val_80'
-'90','val_90'
-'83','val_83'
-'9','val_9'
-'97','val_97'
-'84','val_84'
-'90','val_90'
-'97','val_97'
-21 rows selected 
->>>  
->>>  SELECT * FROM src WHERE key >= '9';
-'key','value'
-'98','val_98'
-'92','val_92'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'90','val_90'
-'95','val_95'
-'90','val_90'
-'9','val_9'
-'97','val_97'
-'90','val_90'
-'97','val_97'
-12 rows selected 
->>>  
->>>  SET hive.exec.post.hooks=;
-No rows affected 
->>>  
->>>  DROP INDEX src_index ON src;
-No rows affected 
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/index_compression.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/index_compression.q.out b/ql/src/test/results/beelinepositive/index_compression.q.out
deleted file mode 100644
index db721dc..0000000
--- a/ql/src/test/results/beelinepositive/index_compression.q.out
+++ /dev/null
@@ -1,176 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/index_compression.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/index_compression.q
->>>  SET hive.exec.compress.result=true;
-No rows affected 
->>>  CREATE INDEX src_index ON TABLE src(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected 
->>>  ALTER INDEX src_index ON src REBUILD;
-No rows affected 
->>>  
->>>  SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected 
->>>  SET hive.optimize.index.filter=true;
-No rows affected 
->>>  SET hive.optimize.index.filter.compact.minsize=0;
-No rows affected 
->>>  
->>>  -- automatic indexing
->>>  EXPLAIN SELECT key, value FROM src WHERE key > 80 AND key < 100 ORDER BY key;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_TABLE_OR_COL value))) (TOK_WHERE (AND (> (TOK_TABLE_OR_COL key) 80) (< (TOK_TABLE_OR_COL key) 100))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-3 is a root stage'
-'  Stage-8 depends on stages: Stage-3 , consists of Stage-5, Stage-4, Stage-6'
-'  Stage-5'
-'  Stage-2 depends on stages: Stage-5, Stage-4, Stage-7'
-'  Stage-1 depends on stages: Stage-2'
-'  Stage-4'
-'  Stage-6'
-'  Stage-7 depends on stages: Stage-6'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-3'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        index_compression__src_src_index__ '
-'          TableScan'
-'            alias: index_compression__src_src_index__'
-'            filterExpr:'
-'                expr: ((key > 80.0) and (key < 100.0))'
-'                type: boolean'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key > 80.0) and (key < 100.0))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: _bucketname'
-'                      type: string'
-'                      expr: _offsets'
-'                      type: array<bigint>'
-'                outputColumnNames: _col0, _col1'
-'                File Output Operator'
-'                  compressed: false'
-'                  GlobalTableId: 1'
-'                  table:'
-'                      input format: org.apache.hadoop.mapred.TextInputFormat'
-'                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-8'
-'    Conditional Operator'
-''
-'  Stage: Stage-5'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          destination: file:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-2'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          destination: file:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src '
-'          TableScan'
-'            alias: src'
-'            filterExpr:'
-'                expr: ((key > 80.0) and (key < 100.0))'
-'                type: boolean'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((key > 80.0) and (key < 100.0))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                Reduce Output Operator'
-'                  key expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                  sort order: +'
-'                  tag: -1'
-'                  value expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                        expr: _col1'
-'                        type: string'
-'      Reduce Operator Tree:'
-'        Extract'
-'          File Output Operator'
-'            compressed: false'
-'            GlobalTableId: 0'
-'            table:'
-'                input format: org.apache.hadoop.mapred.TextInputFormat'
-'                output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-4'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        file:!!{hive.exec.scratchdir}!! '
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-6'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        file:!!{hive.exec.scratchdir}!! '
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-7'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          destination: file:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-''
-''
-130 rows selected 
->>>  SELECT key, value FROM src WHERE key > 80 AND key < 100 ORDER BY key;
-'key','value'
-'82','val_82'
-'83','val_83'
-'83','val_83'
-'84','val_84'
-'84','val_84'
-'85','val_85'
-'86','val_86'
-'87','val_87'
-'90','val_90'
-'90','val_90'
-'90','val_90'
-'92','val_92'
-'95','val_95'
-'95','val_95'
-'96','val_96'
-'97','val_97'
-'97','val_97'
-'98','val_98'
-'98','val_98'
-19 rows selected 
->>>  
->>>  DROP INDEX src_index on src;
-No rows affected 
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/index_stale.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/index_stale.q.out b/ql/src/test/results/beelinepositive/index_stale.q.out
deleted file mode 100644
index 708b504..0000000
--- a/ql/src/test/results/beelinepositive/index_stale.q.out
+++ /dev/null
@@ -1,79 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/index_stale.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/index_stale.q
->>>  -- test that stale indexes are not used
->>>  
->>>  CREATE TABLE temp(key STRING, val STRING) STORED AS TEXTFILE;
-No rows affected 
->>>  INSERT OVERWRITE TABLE temp SELECT * FROM src WHERE key < 50;
-'key','value'
-No rows selected 
->>>  
->>>  -- Build an index on temp.
->>>  CREATE INDEX temp_index ON TABLE temp(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected 
->>>  ALTER INDEX temp_index ON temp REBUILD;
-No rows affected 
->>>  
->>>  SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected 
->>>  SET hive.optimize.index.filter=true;
-No rows affected 
->>>  SET hive.optimize.index.filter.compact.minsize=0;
-No rows affected 
->>>  
->>>  -- overwrite temp table so index is out of date
->>>  INSERT OVERWRITE TABLE temp SELECT * FROM src;
-'key','value'
-No rows selected 
->>>  
->>>  -- should return correct results bypassing index
->>>  EXPLAIN SELECT * FROM temp WHERE key  = 86;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME temp))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (= (TOK_TABLE_OR_COL key) 86))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-1 is a root stage'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        temp '
-'          TableScan'
-'            alias: temp'
-'            filterExpr:'
-'                expr: (key = 86.0)'
-'                type: boolean'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key = 86.0)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: val'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                File Output Operator'
-'                  compressed: false'
-'                  GlobalTableId: 0'
-'                  table:'
-'                      input format: org.apache.hadoop.mapred.TextInputFormat'
-'                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-''
-''
-40 rows selected 
->>>  SELECT * FROM temp WHERE key  = 86;
-'key','val'
-'86','val_86'
-1 row selected 
->>>  DROP table temp;
-No rows affected 
->>>  !record