You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by gu...@apache.org on 2017/02/03 21:50:23 UTC
[10/51] [partial] hive git commit: HIVE-15790: Remove unused beeline
golden files (Gunther Hagleitner, reviewed by Sergey Shelukhin)
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/index_auto_partitioned.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/index_auto_partitioned.q.out b/ql/src/test/results/beelinepositive/index_auto_partitioned.q.out
deleted file mode 100644
index 1b7a226..0000000
--- a/ql/src/test/results/beelinepositive/index_auto_partitioned.q.out
+++ /dev/null
@@ -1,157 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/index_auto_partitioned.q.raw". Enter "record" with no arguments to stop it.
->>> !run !!{qFileDirectory}!!/index_auto_partitioned.q
->>> -- test automatic use of index on table with partitions
->>> CREATE INDEX src_part_index ON TABLE srcpart(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected
->>> ALTER INDEX src_part_index ON srcpart REBUILD;
-No rows affected
->>>
->>> SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected
->>> SET hive.optimize.index.filter=true;
-No rows affected
->>> SET hive.optimize.index.filter.compact.minsize=0;
-No rows affected
->>>
->>> EXPLAIN SELECT key, value FROM srcpart WHERE key=86 AND ds='2008-04-09' ORDER BY key;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME srcpart))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_TABLE_OR_COL value))) (TOK_WHERE (AND (= (TOK_TABLE_OR_COL key) 86) (= (TOK_TABLE_OR_COL ds) '2008-04-09'))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)))))'
-''
-'STAGE DEPENDENCIES:'
-' Stage-3 is a root stage'
-' Stage-8 depends on stages: Stage-3 , consists of Stage-5, Stage-4, Stage-6'
-' Stage-5'
-' Stage-2 depends on stages: Stage-5, Stage-4, Stage-7'
-' Stage-1 depends on stages: Stage-2'
-' Stage-4'
-' Stage-6'
-' Stage-7 depends on stages: Stage-6'
-' Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-' Stage: Stage-3'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' index_auto_partitioned__srcpart_src_part_index__ '
-' TableScan'
-' alias: index_auto_partitioned__srcpart_src_part_index__'
-' filterExpr:'
-' expr: ((key = 86.0) and (ds = '2008-04-09'))'
-' type: boolean'
-' Filter Operator'
-' predicate:'
-' expr: (key = 86.0)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: _bucketname'
-' type: string'
-' expr: _offsets'
-' type: array<bigint>'
-' outputColumnNames: _col0, _col1'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 1'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-8'
-' Conditional Operator'
-''
-' Stage: Stage-5'
-' Move Operator'
-' files:'
-' hdfs directory: true'
-' destination: file:!!{hive.exec.scratchdir}!!'
-''
-' Stage: Stage-2'
-' Move Operator'
-' files:'
-' hdfs directory: true'
-' destination: file:!!{hive.exec.scratchdir}!!'
-''
-' Stage: Stage-1'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' srcpart '
-' TableScan'
-' alias: srcpart'
-' filterExpr:'
-' expr: ((key = 86.0) and (ds = '2008-04-09'))'
-' type: boolean'
-' Filter Operator'
-' predicate:'
-' expr: (key = 86.0)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' tag: -1'
-' value expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: string'
-' Reduce Operator Tree:'
-' Extract'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-4'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' file:!!{hive.exec.scratchdir}!! '
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-6'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' file:!!{hive.exec.scratchdir}!! '
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-7'
-' Move Operator'
-' files:'
-' hdfs directory: true'
-' destination: file:!!{hive.exec.scratchdir}!!'
-''
-' Stage: Stage-0'
-' Fetch Operator'
-' limit: -1'
-''
-''
-130 rows selected
->>> SELECT key, value FROM srcpart WHERE key=86 AND ds='2008-04-09' ORDER BY key;
-'key','value'
-'86','val_86'
-'86','val_86'
-2 rows selected
->>>
->>> DROP INDEX src_part_index ON srcpart;
-No rows affected
->>> !record
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/index_auto_self_join.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/index_auto_self_join.q.out b/ql/src/test/results/beelinepositive/index_auto_self_join.q.out
deleted file mode 100644
index ec7e7f7..0000000
--- a/ql/src/test/results/beelinepositive/index_auto_self_join.q.out
+++ /dev/null
@@ -1,445 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/index_auto_self_join.q.raw". Enter "record" with no arguments to stop it.
->>> !run !!{qFileDirectory}!!/index_auto_self_join.q
->>> -- try the query without indexing, with manual indexing, and with automatic indexing
->>>
->>> EXPLAIN SELECT a.key, b.key FROM src a JOIN src b ON (a.value = b.value) WHERE a.key > 80 AND a.key < 100 AND b.key > 70 AND b.key < 90 ORDER BY a.key;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_TABREF (TOK_TABNAME src) a) (TOK_TABREF (TOK_TABNAME src) b) (= (. (TOK_TABLE_OR_COL a) value) (. (TOK_TABLE_OR_COL b) value)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL a) key)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL b) key))) (TOK_WHERE (AND (AND (AND (> (. (TOK_TABLE_OR_COL a) key) 80) (< (. (TOK_TABLE_OR_COL a) key) 100)) (> (. (TOK_TABLE_OR_COL b) key) 70)) (< (. (TOK_TABLE_OR_COL b) key) 90))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (. (TOK_TABLE_OR_COL a) key)))))'
-''
-'STAGE DEPENDENCIES:'
-' Stage-1 is a root stage'
-' Stage-2 depends on stages: Stage-1'
-' Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-' Stage: Stage-1'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' a '
-' TableScan'
-' alias: a'
-' Filter Operator'
-' predicate:'
-' expr: ((key > 80.0) and (key < 100.0))'
-' type: boolean'
-' Reduce Output Operator'
-' key expressions:'
-' expr: value'
-' type: string'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: value'
-' type: string'
-' tag: 0'
-' value expressions:'
-' expr: key'
-' type: string'
-' b '
-' TableScan'
-' alias: b'
-' Filter Operator'
-' predicate:'
-' expr: ((key > 70.0) and (key < 90.0))'
-' type: boolean'
-' Reduce Output Operator'
-' key expressions:'
-' expr: value'
-' type: string'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: value'
-' type: string'
-' tag: 1'
-' value expressions:'
-' expr: key'
-' type: string'
-' Reduce Operator Tree:'
-' Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' condition expressions:'
-' 0 {VALUE._col0}'
-' 1 {VALUE._col0}'
-' handleSkewJoin: false'
-' outputColumnNames: _col0, _col4'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col4'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-''
-' Stage: Stage-2'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' file:!!{hive.exec.scratchdir}!! '
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' tag: -1'
-' value expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: string'
-' Reduce Operator Tree:'
-' Extract'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-0'
-' Fetch Operator'
-' limit: -1'
-''
-''
-102 rows selected
->>> SELECT a.key, b.key FROM src a JOIN src b ON (a.value = b.value) WHERE a.key > 80 AND a.key < 100 AND b.key > 70 AND b.key < 90 ORDER BY a.key;
-'key','key'
-'82','82'
-'83','83'
-'83','83'
-'83','83'
-'83','83'
-'84','84'
-'84','84'
-'84','84'
-'84','84'
-'85','85'
-'86','86'
-'87','87'
-12 rows selected
->>>
->>> CREATE INDEX src_index ON TABLE src(key) as 'BITMAP' WITH DEFERRED REBUILD;
-No rows affected
->>> ALTER INDEX src_index ON src REBUILD;
-No rows affected
->>>
->>> SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected
->>> SET hive.optimize.index.filter=true;
-No rows affected
->>> SET hive.optimize.index.filter.compact.minsize=0;
-No rows affected
->>>
->>> EXPLAIN SELECT a.key, b.key FROM src a JOIN src b ON (a.value = b.value) WHERE a.key > 80 AND a.key < 100 AND b.key > 70 AND b.key < 90 ORDER BY a.key;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_TABREF (TOK_TABNAME src) a) (TOK_TABREF (TOK_TABNAME src) b) (= (. (TOK_TABLE_OR_COL a) value) (. (TOK_TABLE_OR_COL b) value)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL a) key)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL b) key))) (TOK_WHERE (AND (AND (AND (> (. (TOK_TABLE_OR_COL a) key) 80) (< (. (TOK_TABLE_OR_COL a) key) 100)) (> (. (TOK_TABLE_OR_COL b) key) 70)) (< (. (TOK_TABLE_OR_COL b) key) 90))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (. (TOK_TABLE_OR_COL a) key)))))'
-''
-'STAGE DEPENDENCIES:'
-' Stage-5 is a root stage'
-' Stage-4 depends on stages: Stage-5'
-' Stage-1 depends on stages: Stage-4, Stage-6'
-' Stage-2 depends on stages: Stage-1'
-' Stage-7 is a root stage'
-' Stage-6 depends on stages: Stage-7'
-' Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-' Stage: Stage-5'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' tmp_index:ind0:index_auto_self_join__src_src_index__ '
-' TableScan'
-' alias: index_auto_self_join__src_src_index__'
-' filterExpr:'
-' expr: (((key > 70.0) and (key < 90.0)) and (not EWAH_BITMAP_EMPTY(_bitmaps)))'
-' type: boolean'
-' Filter Operator'
-' predicate:'
-' expr: (((key > 70.0) and (key < 90.0)) and (not EWAH_BITMAP_EMPTY(_bitmaps)))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: _bucketname'
-' type: string'
-' expr: _offset'
-' type: bigint'
-' expr: _bitmaps'
-' type: array<bigint>'
-' outputColumnNames: _col1, _col2, _col3'
-' Select Operator'
-' expressions:'
-' expr: _col1'
-' type: string'
-' expr: _col2'
-' type: bigint'
-' outputColumnNames: _col0, _col1'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: bigint'
-' outputColumnNames: _col0, _col1'
-' Group By Operator'
-' aggregations:'
-' expr: collect_set(_col1)'
-' bucketGroup: false'
-' keys:'
-' expr: _col0'
-' type: string'
-' mode: hash'
-' outputColumnNames: _col0, _col1'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: _col0'
-' type: string'
-' tag: -1'
-' value expressions:'
-' expr: _col1'
-' type: array<bigint>'
-' Reduce Operator Tree:'
-' Group By Operator'
-' aggregations:'
-' expr: collect_set(VALUE._col0)'
-' bucketGroup: false'
-' keys:'
-' expr: KEY._col0'
-' type: string'
-' mode: mergepartial'
-' outputColumnNames: _col0, _col1'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: array<bigint>'
-' outputColumnNames: _col0, _col1'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 1'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-4'
-' Move Operator'
-' files:'
-' hdfs directory: true'
-' destination: file:!!{hive.exec.scratchdir}!!'
-''
-' Stage: Stage-1'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' a '
-' TableScan'
-' alias: a'
-' filterExpr:'
-' expr: ((key > 80.0) and (key < 100.0))'
-' type: boolean'
-' Filter Operator'
-' predicate:'
-' expr: ((key > 80.0) and (key < 100.0))'
-' type: boolean'
-' Reduce Output Operator'
-' key expressions:'
-' expr: value'
-' type: string'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: value'
-' type: string'
-' tag: 0'
-' value expressions:'
-' expr: key'
-' type: string'
-' b '
-' TableScan'
-' alias: b'
-' filterExpr:'
-' expr: ((key > 70.0) and (key < 90.0))'
-' type: boolean'
-' Filter Operator'
-' predicate:'
-' expr: ((key > 70.0) and (key < 90.0))'
-' type: boolean'
-' Reduce Output Operator'
-' key expressions:'
-' expr: value'
-' type: string'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: value'
-' type: string'
-' tag: 1'
-' value expressions:'
-' expr: key'
-' type: string'
-' Reduce Operator Tree:'
-' Join Operator'
-' condition map:'
-' Inner Join 0 to 1'
-' condition expressions:'
-' 0 {VALUE._col0}'
-' 1 {VALUE._col0}'
-' handleSkewJoin: false'
-' outputColumnNames: _col0, _col4'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col4'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-''
-' Stage: Stage-2'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' file:!!{hive.exec.scratchdir}!! '
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' tag: -1'
-' value expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: string'
-' Reduce Operator Tree:'
-' Extract'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-7'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' tmp_index:ind0:index_auto_self_join__src_src_index__ '
-' TableScan'
-' alias: index_auto_self_join__src_src_index__'
-' filterExpr:'
-' expr: (((key > 80.0) and (key < 100.0)) and (not EWAH_BITMAP_EMPTY(_bitmaps)))'
-' type: boolean'
-' Filter Operator'
-' predicate:'
-' expr: (((key > 80.0) and (key < 100.0)) and (not EWAH_BITMAP_EMPTY(_bitmaps)))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: _bucketname'
-' type: string'
-' expr: _offset'
-' type: bigint'
-' expr: _bitmaps'
-' type: array<bigint>'
-' outputColumnNames: _col1, _col2, _col3'
-' Select Operator'
-' expressions:'
-' expr: _col1'
-' type: string'
-' expr: _col2'
-' type: bigint'
-' outputColumnNames: _col0, _col1'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: bigint'
-' outputColumnNames: _col0, _col1'
-' Group By Operator'
-' aggregations:'
-' expr: collect_set(_col1)'
-' bucketGroup: false'
-' keys:'
-' expr: _col0'
-' type: string'
-' mode: hash'
-' outputColumnNames: _col0, _col1'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: _col0'
-' type: string'
-' tag: -1'
-' value expressions:'
-' expr: _col1'
-' type: array<bigint>'
-' Reduce Operator Tree:'
-' Group By Operator'
-' aggregations:'
-' expr: collect_set(VALUE._col0)'
-' bucketGroup: false'
-' keys:'
-' expr: KEY._col0'
-' type: string'
-' mode: mergepartial'
-' outputColumnNames: _col0, _col1'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: array<bigint>'
-' outputColumnNames: _col0, _col1'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 1'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-6'
-' Move Operator'
-' files:'
-' hdfs directory: true'
-' destination: file:!!{hive.exec.scratchdir}!!'
-''
-' Stage: Stage-0'
-' Fetch Operator'
-' limit: -1'
-''
-''
-286 rows selected
->>> SELECT a.key, b.key FROM src a JOIN src b ON (a.value = b.value) WHERE a.key > 80 AND a.key < 100 AND b.key > 70 AND b.key < 90 ORDER BY a.key;
-'key','key'
-'82','82'
-'83','83'
-'83','83'
-'83','83'
-'83','83'
-'84','84'
-'84','84'
-'84','84'
-'84','84'
-'85','85'
-'86','86'
-'87','87'
-12 rows selected
->>>
->>> DROP INDEX src_index on src;
-No rows affected
->>> !record
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/index_auto_unused.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/index_auto_unused.q.out b/ql/src/test/results/beelinepositive/index_auto_unused.q.out
deleted file mode 100644
index cc2a607..0000000
--- a/ql/src/test/results/beelinepositive/index_auto_unused.q.out
+++ /dev/null
@@ -1,484 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/index_auto_unused.q.raw". Enter "record" with no arguments to stop it.
->>> !run !!{qFileDirectory}!!/index_auto_unused.q
->>> -- test cases where the index should not be used automatically
->>>
->>> CREATE INDEX src_index ON TABLE src(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected
->>> ALTER INDEX src_index ON src REBUILD;
-No rows affected
->>>
->>> SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected
->>> SET hive.optimize.index.filter=true;
-No rows affected
->>> SET hive.optimize.index.filter.compact.minsize=5368709120;
-No rows affected
->>> SET hive.optimize.index.filter.compact.maxsize=-1;
-No rows affected
->>>
->>> -- min size too large (src is less than 5G)
->>> EXPLAIN SELECT * FROM src WHERE key > 80 AND key < 100 ORDER BY key;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (AND (> (TOK_TABLE_OR_COL key) 80) (< (TOK_TABLE_OR_COL key) 100))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)))))'
-''
-'STAGE DEPENDENCIES:'
-' Stage-1 is a root stage'
-' Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-' Stage: Stage-1'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src '
-' TableScan'
-' alias: src'
-' filterExpr:'
-' expr: ((key > 80.0) and (key < 100.0))'
-' type: boolean'
-' Filter Operator'
-' predicate:'
-' expr: ((key > 80.0) and (key < 100.0))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' tag: -1'
-' value expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: string'
-' Reduce Operator Tree:'
-' Extract'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-0'
-' Fetch Operator'
-' limit: -1'
-''
-''
-53 rows selected
->>> SELECT * FROM src WHERE key > 80 AND key < 100 ORDER BY key;
-'key','value'
-'82','val_82'
-'83','val_83'
-'83','val_83'
-'84','val_84'
-'84','val_84'
-'85','val_85'
-'86','val_86'
-'87','val_87'
-'90','val_90'
-'90','val_90'
-'90','val_90'
-'92','val_92'
-'95','val_95'
-'95','val_95'
-'96','val_96'
-'97','val_97'
-'97','val_97'
-'98','val_98'
-'98','val_98'
-19 rows selected
->>>
->>> SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected
->>> SET hive.optimize.index.filter=true;
-No rows affected
->>> SET hive.optimize.index.filter.compact.minsize=0;
-No rows affected
->>> SET hive.optimize.index.filter.compact.maxsize=1;
-No rows affected
->>>
->>> -- max size too small
->>> EXPLAIN SELECT * FROM src WHERE key > 80 AND key < 100 ORDER BY key;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (AND (> (TOK_TABLE_OR_COL key) 80) (< (TOK_TABLE_OR_COL key) 100))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)))))'
-''
-'STAGE DEPENDENCIES:'
-' Stage-1 is a root stage'
-' Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-' Stage: Stage-1'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src '
-' TableScan'
-' alias: src'
-' filterExpr:'
-' expr: ((key > 80.0) and (key < 100.0))'
-' type: boolean'
-' Filter Operator'
-' predicate:'
-' expr: ((key > 80.0) and (key < 100.0))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' tag: -1'
-' value expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: string'
-' Reduce Operator Tree:'
-' Extract'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-0'
-' Fetch Operator'
-' limit: -1'
-''
-''
-53 rows selected
->>> SELECT * FROM src WHERE key > 80 AND key < 100 ORDER BY key;
-'key','value'
-'82','val_82'
-'83','val_83'
-'83','val_83'
-'84','val_84'
-'84','val_84'
-'85','val_85'
-'86','val_86'
-'87','val_87'
-'90','val_90'
-'90','val_90'
-'90','val_90'
-'92','val_92'
-'95','val_95'
-'95','val_95'
-'96','val_96'
-'97','val_97'
-'97','val_97'
-'98','val_98'
-'98','val_98'
-19 rows selected
->>>
->>> SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected
->>> SET hive.optimize.index.filter=true;
-No rows affected
->>> SET hive.optimize.index.filter.compact.minsize=0;
-No rows affected
->>> SET hive.optimize.index.filter.compact.maxsize=-1;
-No rows affected
->>>
->>> -- OR predicate not supported by compact indexes
->>> EXPLAIN SELECT * FROM src WHERE key < 10 OR key > 480 ORDER BY key;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (OR (< (TOK_TABLE_OR_COL key) 10) (> (TOK_TABLE_OR_COL key) 480))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)))))'
-''
-'STAGE DEPENDENCIES:'
-' Stage-1 is a root stage'
-' Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-' Stage: Stage-1'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src '
-' TableScan'
-' alias: src'
-' filterExpr:'
-' expr: ((key < 10.0) or (key > 480.0))'
-' type: boolean'
-' Filter Operator'
-' predicate:'
-' expr: ((key < 10.0) or (key > 480.0))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' tag: -1'
-' value expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: string'
-' Reduce Operator Tree:'
-' Extract'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-0'
-' Fetch Operator'
-' limit: -1'
-''
-''
-53 rows selected
->>> SELECT * FROM src WHERE key < 10 OR key > 480 ORDER BY key;
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-'2','val_2'
-'4','val_4'
-'481','val_481'
-'482','val_482'
-'483','val_483'
-'484','val_484'
-'485','val_485'
-'487','val_487'
-'489','val_489'
-'489','val_489'
-'489','val_489'
-'489','val_489'
-'490','val_490'
-'491','val_491'
-'492','val_492'
-'492','val_492'
-'493','val_493'
-'494','val_494'
-'495','val_495'
-'496','val_496'
-'497','val_497'
-'498','val_498'
-'498','val_498'
-'498','val_498'
-'5','val_5'
-'5','val_5'
-'5','val_5'
-'8','val_8'
-'9','val_9'
-32 rows selected
->>>
->>> SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected
->>> SET hive.optimize.index.filter=true;
-No rows affected
->>> SET hive.optimize.index.filter.compact.minsize=0;
-No rows affected
->>> SET hive.optimize.index.filter.compact.maxsize=-1;
-No rows affected
->>>
->>> -- columns are not covered by indexes
->>> DROP INDEX src_index on src;
-No rows affected
->>> CREATE INDEX src_val_index ON TABLE src(value) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected
->>> ALTER INDEX src_val_index ON src REBUILD;
-No rows affected
->>>
->>> EXPLAIN SELECT * FROM src WHERE key > 80 AND key < 100 ORDER BY key;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (AND (> (TOK_TABLE_OR_COL key) 80) (< (TOK_TABLE_OR_COL key) 100))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)))))'
-''
-'STAGE DEPENDENCIES:'
-' Stage-1 is a root stage'
-' Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-' Stage: Stage-1'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src '
-' TableScan'
-' alias: src'
-' filterExpr:'
-' expr: ((key > 80.0) and (key < 100.0))'
-' type: boolean'
-' Filter Operator'
-' predicate:'
-' expr: ((key > 80.0) and (key < 100.0))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' tag: -1'
-' value expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: string'
-' Reduce Operator Tree:'
-' Extract'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-0'
-' Fetch Operator'
-' limit: -1'
-''
-''
-53 rows selected
->>> SELECT * FROM src WHERE key > 80 AND key < 100 ORDER BY key;
-'key','value'
-'82','val_82'
-'83','val_83'
-'83','val_83'
-'84','val_84'
-'84','val_84'
-'85','val_85'
-'86','val_86'
-'87','val_87'
-'90','val_90'
-'90','val_90'
-'90','val_90'
-'92','val_92'
-'95','val_95'
-'95','val_95'
-'96','val_96'
-'97','val_97'
-'97','val_97'
-'98','val_98'
-'98','val_98'
-19 rows selected
->>>
->>> DROP INDEX src_val_index on src;
-No rows affected
->>>
->>> SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected
->>> SET hive.optimize.index.filter=true;
-No rows affected
->>> SET hive.optimize.index.filter.compact.minsize=0;
-No rows affected
->>> SET hive.optimize.index.filter.compact.maxsize=-1;
-No rows affected
->>>
->>> -- required partitions have not been built yet
->>> CREATE INDEX src_part_index ON TABLE srcpart(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected
->>> ALTER INDEX src_part_index ON srcpart PARTITION (ds='2008-04-08', hr=11) REBUILD;
-No rows affected
->>>
->>> EXPLAIN SELECT * FROM srcpart WHERE ds='2008-04-09' AND hr=12 AND key < 10 ORDER BY key;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME srcpart))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (AND (AND (= (TOK_TABLE_OR_COL ds) '2008-04-09') (= (TOK_TABLE_OR_COL hr) 12)) (< (TOK_TABLE_OR_COL key) 10))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)))))'
-''
-'STAGE DEPENDENCIES:'
-' Stage-1 is a root stage'
-' Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-' Stage: Stage-1'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' srcpart '
-' TableScan'
-' alias: srcpart'
-' filterExpr:'
-' expr: (((ds = '2008-04-09') and (hr = 12.0)) and (key < 10.0))'
-' type: boolean'
-' Filter Operator'
-' predicate:'
-' expr: (key < 10.0)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' expr: ds'
-' type: string'
-' expr: hr'
-' type: string'
-' outputColumnNames: _col0, _col1, _col2, _col3'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' tag: -1'
-' value expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: string'
-' expr: _col2'
-' type: string'
-' expr: _col3'
-' type: string'
-' Reduce Operator Tree:'
-' Extract'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-0'
-' Fetch Operator'
-' limit: -1'
-''
-''
-61 rows selected
->>> SELECT * FROM srcpart WHERE ds='2008-04-09' AND hr=12 AND key < 10 ORDER BY key;
-'key','value','ds','hr'
-'0','val_0','2008-04-09','12'
-'0','val_0','2008-04-09','12'
-'0','val_0','2008-04-09','12'
-'2','val_2','2008-04-09','12'
-'4','val_4','2008-04-09','12'
-'5','val_5','2008-04-09','12'
-'5','val_5','2008-04-09','12'
-'5','val_5','2008-04-09','12'
-'8','val_8','2008-04-09','12'
-'9','val_9','2008-04-09','12'
-10 rows selected
->>>
->>> DROP INDEX src_part_index on srcpart;
-No rows affected
->>> !record
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/index_auto_update.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/index_auto_update.q.out b/ql/src/test/results/beelinepositive/index_auto_update.q.out
deleted file mode 100644
index 3330998..0000000
--- a/ql/src/test/results/beelinepositive/index_auto_update.q.out
+++ /dev/null
@@ -1,352 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/index_auto_update.q.raw". Enter "record" with no arguments to stop it.
->>> !run !!{qFileDirectory}!!/index_auto_update.q
->>> -- Test if index is actually being used.
->>>
->>> -- Create temp, and populate it with some values in src.
->>> CREATE TABLE temp(key STRING, val STRING) STORED AS TEXTFILE;
-No rows affected
->>> INSERT OVERWRITE TABLE temp SELECT * FROM src WHERE key < 50;
-'key','value'
-No rows selected
->>>
->>> -- Build an index on temp.
->>> CREATE INDEX temp_index ON TABLE temp(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected
->>> ALTER INDEX temp_index ON temp REBUILD;
-No rows affected
->>>
->>> SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected
->>> SET hive.optimize.index.filter=true;
-No rows affected
->>> SET hive.optimize.index.autoupdate=true;
-No rows affected
->>> SET hive.optimize.index.filter.compact.minsize=0;
-No rows affected
->>>
->>> -- overwrite temp table so index is out of date
->>> EXPLAIN INSERT OVERWRITE TABLE temp SELECT * FROM src;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME temp))) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF))))'
-''
-'STAGE DEPENDENCIES:'
-' Stage-4 is a root stage'
-' Stage-10 depends on stages: Stage-4 , consists of Stage-7, Stage-6, Stage-8'
-' Stage-7'
-' Stage-0 depends on stages: Stage-7, Stage-6, Stage-9'
-' Stage-1 depends on stages: Stage-0'
-' Stage-0 depends on stages: Stage-1'
-' Stage-2 depends on stages: Stage-0'
-' null depends on stages: Stage-1'
-' Stage-3 depends on stages: Stage-1'
-' Stage-5 depends on stages: Stage-0'
-' Stage-6'
-' Stage-8'
-' Stage-9 depends on stages: Stage-8'
-''
-'STAGE PLANS:'
-' Stage: Stage-4'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src '
-' TableScan'
-' alias: src'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 1'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-' name: index_auto_update.temp'
-''
-' Stage: Stage-10'
-' Conditional Operator'
-''
-' Stage: Stage-7'
-' Move Operator'
-' files:'
-' hdfs directory: true'
-' destination: pfile:!!{hive.exec.scratchdir}!!'
-''
-' Stage: Stage-0'
-' Move Operator'
-' tables:'
-' replace: true'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-' name: index_auto_update.temp'
-''
-' Stage: Stage-1'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' temp '
-' TableScan'
-' alias: temp'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: INPUT__FILE__NAME'
-' type: string'
-' expr: BLOCK__OFFSET__INSIDE__FILE'
-' type: bigint'
-' outputColumnNames: key, INPUT__FILE__NAME, BLOCK__OFFSET__INSIDE__FILE'
-' Group By Operator'
-' aggregations:'
-' expr: collect_set(BLOCK__OFFSET__INSIDE__FILE)'
-' bucketGroup: false'
-' keys:'
-' expr: key'
-' type: string'
-' expr: INPUT__FILE__NAME'
-' type: string'
-' mode: hash'
-' outputColumnNames: _col0, _col1, _col2'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: string'
-' sort order: ++'
-' Map-reduce partition columns:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: string'
-' tag: -1'
-' value expressions:'
-' expr: _col2'
-' type: array<bigint>'
-' Reduce Operator Tree:'
-' Group By Operator'
-' aggregations:'
-' expr: collect_set(VALUE._col0)'
-' bucketGroup: false'
-' keys:'
-' expr: KEY._col0'
-' type: string'
-' expr: KEY._col1'
-' type: string'
-' mode: mergepartial'
-' outputColumnNames: _col0, _col1, _col2'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: string'
-' expr: _col2'
-' type: array<bigint>'
-' outputColumnNames: _col0, _col1, _col2'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 1'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-' name: index_auto_update.index_auto_update__temp_temp_index__'
-''
-' Stage: Stage-0'
-' Move Operator'
-' tables:'
-' replace: true'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-' name: index_auto_update.index_auto_update__temp_temp_index__'
-''
-' Stage: Stage-2'
-' Stats-Aggr Operator'
-''
-' Stage: null'
-''
-' Stage: Stage-3'
-''
-' Stage: Stage-5'
-' Stats-Aggr Operator'
-''
-' Stage: Stage-6'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' pfile:!!{hive.exec.scratchdir}!! '
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-' name: index_auto_update.temp'
-''
-' Stage: Stage-8'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' pfile:!!{hive.exec.scratchdir}!! '
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-' name: index_auto_update.temp'
-''
-' Stage: Stage-9'
-' Move Operator'
-' files:'
-' hdfs directory: true'
-' destination: pfile:!!{hive.exec.scratchdir}!!'
-''
-''
-185 rows selected
->>> INSERT OVERWRITE TABLE temp SELECT * FROM src;
-'key','value'
-No rows selected
->>>
->>> -- query should return indexed values
->>> EXPLAIN SELECT * FROM temp WHERE key = 86;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME temp))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (= (TOK_TABLE_OR_COL key) 86))))'
-''
-'STAGE DEPENDENCIES:'
-' Stage-3 is a root stage'
-' Stage-8 depends on stages: Stage-3 , consists of Stage-5, Stage-4, Stage-6'
-' Stage-5'
-' Stage-2 depends on stages: Stage-5, Stage-4, Stage-7'
-' Stage-1 depends on stages: Stage-2'
-' Stage-4'
-' Stage-6'
-' Stage-7 depends on stages: Stage-6'
-' Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-' Stage: Stage-3'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' index_auto_update__temp_temp_index__ '
-' TableScan'
-' alias: index_auto_update__temp_temp_index__'
-' filterExpr:'
-' expr: (key = 86.0)'
-' type: boolean'
-' Filter Operator'
-' predicate:'
-' expr: (key = 86.0)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: _bucketname'
-' type: string'
-' expr: _offsets'
-' type: array<bigint>'
-' outputColumnNames: _col0, _col1'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 1'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-8'
-' Conditional Operator'
-''
-' Stage: Stage-5'
-' Move Operator'
-' files:'
-' hdfs directory: true'
-' destination: file:!!{hive.exec.scratchdir}!!'
-''
-' Stage: Stage-2'
-' Move Operator'
-' files:'
-' hdfs directory: true'
-' destination: file:!!{hive.exec.scratchdir}!!'
-''
-' Stage: Stage-1'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' temp '
-' TableScan'
-' alias: temp'
-' filterExpr:'
-' expr: (key = 86.0)'
-' type: boolean'
-' Filter Operator'
-' predicate:'
-' expr: (key = 86.0)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: val'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-4'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' file:!!{hive.exec.scratchdir}!! '
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-6'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' file:!!{hive.exec.scratchdir}!! '
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-7'
-' Move Operator'
-' files:'
-' hdfs directory: true'
-' destination: file:!!{hive.exec.scratchdir}!!'
-''
-' Stage: Stage-0'
-' Fetch Operator'
-' limit: -1'
-''
-''
-117 rows selected
->>> SELECT * FROM temp WHERE key = 86;
-'key','val'
-'86','val_86'
-1 row selected
->>>
->>> SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected
->>> SET hive.optimize.index.filter=false;
-No rows affected
->>> DROP table temp;
-No rows affected
->>> !record
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/index_bitmap_auto_partitioned.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/index_bitmap_auto_partitioned.q.out b/ql/src/test/results/beelinepositive/index_bitmap_auto_partitioned.q.out
deleted file mode 100644
index 3dcd175..0000000
--- a/ql/src/test/results/beelinepositive/index_bitmap_auto_partitioned.q.out
+++ /dev/null
@@ -1,167 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/index_bitmap_auto_partitioned.q.raw". Enter "record" with no arguments to stop it.
->>> !run !!{qFileDirectory}!!/index_bitmap_auto_partitioned.q
->>> -- test automatic use of index on table with partitions
->>> CREATE INDEX src_part_index ON TABLE srcpart(key) as 'BITMAP' WITH DEFERRED REBUILD;
-No rows affected
->>> ALTER INDEX src_part_index ON srcpart REBUILD;
-No rows affected
->>>
->>> SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected
->>> SET hive.optimize.index.filter=true;
-No rows affected
->>>
->>> EXPLAIN SELECT key, value FROM srcpart WHERE key=86 AND ds='2008-04-09' ORDER BY key;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME srcpart))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_TABLE_OR_COL value))) (TOK_WHERE (AND (= (TOK_TABLE_OR_COL key) 86) (= (TOK_TABLE_OR_COL ds) '2008-04-09'))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)))))'
-''
-'STAGE DEPENDENCIES:'
-' Stage-3 is a root stage'
-' Stage-2 depends on stages: Stage-3'
-' Stage-1 depends on stages: Stage-2'
-' Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-' Stage: Stage-3'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' tmp_index:ind0:index_bitmap_auto_partitioned__srcpart_src_part_index__ '
-' TableScan'
-' alias: index_bitmap_auto_partitioned__srcpart_src_part_index__'
-' filterExpr:'
-' expr: ((key = 86.0) and (not EWAH_BITMAP_EMPTY(_bitmaps)))'
-' type: boolean'
-' Filter Operator'
-' predicate:'
-' expr: ((key = 86.0) and (not EWAH_BITMAP_EMPTY(_bitmaps)))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: _bucketname'
-' type: string'
-' expr: _offset'
-' type: bigint'
-' expr: _bitmaps'
-' type: array<bigint>'
-' outputColumnNames: _col1, _col2, _col3'
-' Select Operator'
-' expressions:'
-' expr: _col1'
-' type: string'
-' expr: _col2'
-' type: bigint'
-' outputColumnNames: _col0, _col1'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: bigint'
-' outputColumnNames: _col0, _col1'
-' Group By Operator'
-' aggregations:'
-' expr: collect_set(_col1)'
-' bucketGroup: false'
-' keys:'
-' expr: _col0'
-' type: string'
-' mode: hash'
-' outputColumnNames: _col0, _col1'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: _col0'
-' type: string'
-' tag: -1'
-' value expressions:'
-' expr: _col1'
-' type: array<bigint>'
-' Reduce Operator Tree:'
-' Group By Operator'
-' aggregations:'
-' expr: collect_set(VALUE._col0)'
-' bucketGroup: false'
-' keys:'
-' expr: KEY._col0'
-' type: string'
-' mode: mergepartial'
-' outputColumnNames: _col0, _col1'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: array<bigint>'
-' outputColumnNames: _col0, _col1'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 1'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-2'
-' Move Operator'
-' files:'
-' hdfs directory: true'
-' destination: file:!!{hive.exec.scratchdir}!!'
-''
-' Stage: Stage-1'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' srcpart '
-' TableScan'
-' alias: srcpart'
-' filterExpr:'
-' expr: ((key = 86.0) and (ds = '2008-04-09'))'
-' type: boolean'
-' Filter Operator'
-' predicate:'
-' expr: (key = 86.0)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' tag: -1'
-' value expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: string'
-' Reduce Operator Tree:'
-' Extract'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-0'
-' Fetch Operator'
-' limit: -1'
-''
-''
-142 rows selected
->>> SELECT key, value FROM srcpart WHERE key=86 AND ds='2008-04-09' ORDER BY key;
-'key','value'
-'86','val_86'
-'86','val_86'
-2 rows selected
->>>
->>> DROP INDEX src_part_index ON srcpart;
-No rows affected
->>> !record
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/index_bitmap_compression.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/index_bitmap_compression.q.out b/ql/src/test/results/beelinepositive/index_bitmap_compression.q.out
deleted file mode 100644
index 9bf0106..0000000
--- a/ql/src/test/results/beelinepositive/index_bitmap_compression.q.out
+++ /dev/null
@@ -1,188 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/index_bitmap_compression.q.raw". Enter "record" with no arguments to stop it.
->>> !run !!{qFileDirectory}!!/index_bitmap_compression.q
->>> SET hive.exec.compress.result=true;
-No rows affected
->>> CREATE INDEX src_index ON TABLE src(key) as 'BITMAP' WITH DEFERRED REBUILD;
-No rows affected
->>> ALTER INDEX src_index ON src REBUILD;
-No rows affected
->>>
->>> SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected
->>> SET hive.optimize.index.filter=true;
-No rows affected
->>> SET hive.optimize.index.filter.compact.minsize=0;
-No rows affected
->>>
->>> -- automatic indexing
->>> EXPLAIN SELECT key, value FROM src WHERE key > 80 AND key < 100 ORDER BY key;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_TABLE_OR_COL value))) (TOK_WHERE (AND (> (TOK_TABLE_OR_COL key) 80) (< (TOK_TABLE_OR_COL key) 100))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)))))'
-''
-'STAGE DEPENDENCIES:'
-' Stage-3 is a root stage'
-' Stage-2 depends on stages: Stage-3'
-' Stage-1 depends on stages: Stage-2'
-' Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-' Stage: Stage-3'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' tmp_index:ind0:index_bitmap_compression__src_src_index__ '
-' TableScan'
-' alias: index_bitmap_compression__src_src_index__'
-' filterExpr:'
-' expr: (((key > 80.0) and (key < 100.0)) and (not EWAH_BITMAP_EMPTY(_bitmaps)))'
-' type: boolean'
-' Filter Operator'
-' predicate:'
-' expr: (((key > 80.0) and (key < 100.0)) and (not EWAH_BITMAP_EMPTY(_bitmaps)))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: _bucketname'
-' type: string'
-' expr: _offset'
-' type: bigint'
-' expr: _bitmaps'
-' type: array<bigint>'
-' outputColumnNames: _col1, _col2, _col3'
-' Select Operator'
-' expressions:'
-' expr: _col1'
-' type: string'
-' expr: _col2'
-' type: bigint'
-' outputColumnNames: _col0, _col1'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: bigint'
-' outputColumnNames: _col0, _col1'
-' Group By Operator'
-' aggregations:'
-' expr: collect_set(_col1)'
-' bucketGroup: false'
-' keys:'
-' expr: _col0'
-' type: string'
-' mode: hash'
-' outputColumnNames: _col0, _col1'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' Map-reduce partition columns:'
-' expr: _col0'
-' type: string'
-' tag: -1'
-' value expressions:'
-' expr: _col1'
-' type: array<bigint>'
-' Reduce Operator Tree:'
-' Group By Operator'
-' aggregations:'
-' expr: collect_set(VALUE._col0)'
-' bucketGroup: false'
-' keys:'
-' expr: KEY._col0'
-' type: string'
-' mode: mergepartial'
-' outputColumnNames: _col0, _col1'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: array<bigint>'
-' outputColumnNames: _col0, _col1'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 1'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-2'
-' Move Operator'
-' files:'
-' hdfs directory: true'
-' destination: file:!!{hive.exec.scratchdir}!!'
-''
-' Stage: Stage-1'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src '
-' TableScan'
-' alias: src'
-' filterExpr:'
-' expr: ((key > 80.0) and (key < 100.0))'
-' type: boolean'
-' Filter Operator'
-' predicate:'
-' expr: ((key > 80.0) and (key < 100.0))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' tag: -1'
-' value expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: string'
-' Reduce Operator Tree:'
-' Extract'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-0'
-' Fetch Operator'
-' limit: -1'
-''
-''
-142 rows selected
->>> SELECT key, value FROM src WHERE key > 80 AND key < 100 ORDER BY key;
-'key','value'
-'82','val_82'
-'83','val_83'
-'83','val_83'
-'84','val_84'
-'84','val_84'
-'85','val_85'
-'86','val_86'
-'87','val_87'
-'90','val_90'
-'90','val_90'
-'90','val_90'
-'92','val_92'
-'95','val_95'
-'95','val_95'
-'96','val_96'
-'97','val_97'
-'97','val_97'
-'98','val_98'
-'98','val_98'
-19 rows selected
->>>
->>> DROP INDEX src_index on src;
-No rows affected
->>> !record
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/index_compact_binary_search.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/index_compact_binary_search.q.out b/ql/src/test/results/beelinepositive/index_compact_binary_search.q.out
deleted file mode 100644
index e963587..0000000
--- a/ql/src/test/results/beelinepositive/index_compact_binary_search.q.out
+++ /dev/null
@@ -1,489 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/index_compact_binary_search.q.raw". Enter "record" with no arguments to stop it.
->>> !run !!{qFileDirectory}!!/index_compact_binary_search.q
->>> SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected
->>> SET hive.default.fileformat=TextFile;
-No rows affected
->>>
->>> CREATE INDEX src_index ON TABLE src(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected
->>> ALTER INDEX src_index ON src REBUILD;
-No rows affected
->>>
->>> SET hive.optimize.index.filter=true;
-No rows affected
->>> SET hive.optimize.index.filter.compact.minsize=1;
-No rows affected
->>> SET hive.index.compact.binary.search=true;
-No rows affected
->>>
->>> SET hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.VerifyHiveSortedInputFormatUsedHook;
-No rows affected
->>>
->>> SELECT * FROM src WHERE key = '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected
->>>
->>> SELECT * FROM src WHERE key < '1';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected
->>>
->>> SELECT * FROM src WHERE key <= '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected
->>>
->>> SELECT * FROM src WHERE key > '8';
-'key','value'
-'86','val_86'
-'98','val_98'
-'82','val_82'
-'92','val_92'
-'83','val_83'
-'84','val_84'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'85','val_85'
-'87','val_87'
-'90','val_90'
-'95','val_95'
-'80','val_80'
-'90','val_90'
-'83','val_83'
-'9','val_9'
-'97','val_97'
-'84','val_84'
-'90','val_90'
-'97','val_97'
-21 rows selected
->>>
->>> SELECT * FROM src WHERE key >= '9';
-'key','value'
-'98','val_98'
-'92','val_92'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'90','val_90'
-'95','val_95'
-'90','val_90'
-'9','val_9'
-'97','val_97'
-'90','val_90'
-'97','val_97'
-12 rows selected
->>>
->>> SET hive.exec.post.hooks=;
-No rows affected
->>>
->>> DROP INDEX src_index ON src;
-No rows affected
->>>
->>> SET hive.default.fileformat=RCFILE;
-No rows affected
->>>
->>> CREATE INDEX src_index ON TABLE src(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected
->>> ALTER INDEX src_index ON src REBUILD;
-No rows affected
->>>
->>> SET hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.VerifyHiveSortedInputFormatUsedHook;
-No rows affected
->>>
->>> SELECT * FROM src WHERE key = '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected
->>>
->>> SELECT * FROM src WHERE key < '1';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected
->>>
->>> SELECT * FROM src WHERE key <= '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected
->>>
->>> SELECT * FROM src WHERE key > '8';
-'key','value'
-'86','val_86'
-'98','val_98'
-'82','val_82'
-'92','val_92'
-'83','val_83'
-'84','val_84'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'85','val_85'
-'87','val_87'
-'90','val_90'
-'95','val_95'
-'80','val_80'
-'90','val_90'
-'83','val_83'
-'9','val_9'
-'97','val_97'
-'84','val_84'
-'90','val_90'
-'97','val_97'
-21 rows selected
->>>
->>> SELECT * FROM src WHERE key >= '9';
-'key','value'
-'98','val_98'
-'92','val_92'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'90','val_90'
-'95','val_95'
-'90','val_90'
-'9','val_9'
-'97','val_97'
-'90','val_90'
-'97','val_97'
-12 rows selected
->>>
->>> SET hive.exec.post.hooks=;
-No rows affected
->>>
->>> DROP INDEX src_index ON src;
-No rows affected
->>>
->>> SET hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat;
-No rows affected
->>> SET hive.default.fileformat=TextFile;
-No rows affected
->>>
->>> CREATE INDEX src_index ON TABLE src(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected
->>> ALTER INDEX src_index ON src REBUILD;
-No rows affected
->>>
->>> SET hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.VerifyHiveSortedInputFormatUsedHook;
-No rows affected
->>>
->>> SELECT * FROM src WHERE key = '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected
->>>
->>> SELECT * FROM src WHERE key < '1';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected
->>>
->>> SELECT * FROM src WHERE key <= '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected
->>>
->>> SELECT * FROM src WHERE key > '8';
-'key','value'
-'86','val_86'
-'98','val_98'
-'82','val_82'
-'92','val_92'
-'83','val_83'
-'84','val_84'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'85','val_85'
-'87','val_87'
-'90','val_90'
-'95','val_95'
-'80','val_80'
-'90','val_90'
-'83','val_83'
-'9','val_9'
-'97','val_97'
-'84','val_84'
-'90','val_90'
-'97','val_97'
-21 rows selected
->>>
->>> SELECT * FROM src WHERE key >= '9';
-'key','value'
-'98','val_98'
-'92','val_92'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'90','val_90'
-'95','val_95'
-'90','val_90'
-'9','val_9'
-'97','val_97'
-'90','val_90'
-'97','val_97'
-12 rows selected
->>>
->>> SET hive.exec.post.hooks=;
-No rows affected
->>>
->>> DROP INDEX src_index ON src;
-No rows affected
->>>
->>> SET hive.default.fileformat=RCFILE;
-No rows affected
->>>
->>> CREATE INDEX src_index ON TABLE src(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected
->>> ALTER INDEX src_index ON src REBUILD;
-No rows affected
->>>
->>> SET hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.VerifyHiveSortedInputFormatUsedHook;
-No rows affected
->>>
->>> SELECT * FROM src WHERE key = '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected
->>>
->>> SELECT * FROM src WHERE key < '1';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected
->>>
->>> SELECT * FROM src WHERE key <= '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected
->>>
->>> SELECT * FROM src WHERE key > '8';
-'key','value'
-'86','val_86'
-'98','val_98'
-'82','val_82'
-'92','val_92'
-'83','val_83'
-'84','val_84'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'85','val_85'
-'87','val_87'
-'90','val_90'
-'95','val_95'
-'80','val_80'
-'90','val_90'
-'83','val_83'
-'9','val_9'
-'97','val_97'
-'84','val_84'
-'90','val_90'
-'97','val_97'
-21 rows selected
->>>
->>> SELECT * FROM src WHERE key >= '9';
-'key','value'
-'98','val_98'
-'92','val_92'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'90','val_90'
-'95','val_95'
-'90','val_90'
-'9','val_9'
-'97','val_97'
-'90','val_90'
-'97','val_97'
-12 rows selected
->>>
->>> SET hive.exec.post.hooks=;
-No rows affected
->>>
->>> DROP INDEX src_index ON src;
-No rows affected
->>>
->>> SET hive.input.format=org.apache.hadoop.hive.ql.io.BucketizedHiveInputFormat;
-No rows affected
->>> SET hive.default.fileformat=TextFile;
-No rows affected
->>>
->>> CREATE INDEX src_index ON TABLE src(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected
->>> ALTER INDEX src_index ON src REBUILD;
-No rows affected
->>>
->>> SET hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.VerifyHiveSortedInputFormatUsedHook;
-No rows affected
->>>
->>> SELECT * FROM src WHERE key = '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected
->>>
->>> SELECT * FROM src WHERE key < '1';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected
->>>
->>> SELECT * FROM src WHERE key <= '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected
->>>
->>> SELECT * FROM src WHERE key > '8';
-'key','value'
-'86','val_86'
-'98','val_98'
-'82','val_82'
-'92','val_92'
-'83','val_83'
-'84','val_84'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'85','val_85'
-'87','val_87'
-'90','val_90'
-'95','val_95'
-'80','val_80'
-'90','val_90'
-'83','val_83'
-'9','val_9'
-'97','val_97'
-'84','val_84'
-'90','val_90'
-'97','val_97'
-21 rows selected
->>>
->>> SELECT * FROM src WHERE key >= '9';
-'key','value'
-'98','val_98'
-'92','val_92'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'90','val_90'
-'95','val_95'
-'90','val_90'
-'9','val_9'
-'97','val_97'
-'90','val_90'
-'97','val_97'
-12 rows selected
->>>
->>> SET hive.exec.post.hooks=;
-No rows affected
->>>
->>> DROP INDEX src_index ON src;
-No rows affected
->>>
->>> SET hive.default.fileformat=RCFILE;
-No rows affected
->>>
->>> CREATE INDEX src_index ON TABLE src(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected
->>> ALTER INDEX src_index ON src REBUILD;
-No rows affected
->>>
->>> SET hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.VerifyHiveSortedInputFormatUsedHook;
-No rows affected
->>>
->>> SELECT * FROM src WHERE key = '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected
->>>
->>> SELECT * FROM src WHERE key < '1';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected
->>>
->>> SELECT * FROM src WHERE key <= '0';
-'key','value'
-'0','val_0'
-'0','val_0'
-'0','val_0'
-3 rows selected
->>>
->>> SELECT * FROM src WHERE key > '8';
-'key','value'
-'86','val_86'
-'98','val_98'
-'82','val_82'
-'92','val_92'
-'83','val_83'
-'84','val_84'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'85','val_85'
-'87','val_87'
-'90','val_90'
-'95','val_95'
-'80','val_80'
-'90','val_90'
-'83','val_83'
-'9','val_9'
-'97','val_97'
-'84','val_84'
-'90','val_90'
-'97','val_97'
-21 rows selected
->>>
->>> SELECT * FROM src WHERE key >= '9';
-'key','value'
-'98','val_98'
-'92','val_92'
-'96','val_96'
-'95','val_95'
-'98','val_98'
-'90','val_90'
-'95','val_95'
-'90','val_90'
-'9','val_9'
-'97','val_97'
-'90','val_90'
-'97','val_97'
-12 rows selected
->>>
->>> SET hive.exec.post.hooks=;
-No rows affected
->>>
->>> DROP INDEX src_index ON src;
-No rows affected
->>> !record
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/index_compression.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/index_compression.q.out b/ql/src/test/results/beelinepositive/index_compression.q.out
deleted file mode 100644
index db721dc..0000000
--- a/ql/src/test/results/beelinepositive/index_compression.q.out
+++ /dev/null
@@ -1,176 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/index_compression.q.raw". Enter "record" with no arguments to stop it.
->>> !run !!{qFileDirectory}!!/index_compression.q
->>> SET hive.exec.compress.result=true;
-No rows affected
->>> CREATE INDEX src_index ON TABLE src(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected
->>> ALTER INDEX src_index ON src REBUILD;
-No rows affected
->>>
->>> SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected
->>> SET hive.optimize.index.filter=true;
-No rows affected
->>> SET hive.optimize.index.filter.compact.minsize=0;
-No rows affected
->>>
->>> -- automatic indexing
->>> EXPLAIN SELECT key, value FROM src WHERE key > 80 AND key < 100 ORDER BY key;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_TABLE_OR_COL value))) (TOK_WHERE (AND (> (TOK_TABLE_OR_COL key) 80) (< (TOK_TABLE_OR_COL key) 100))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)))))'
-''
-'STAGE DEPENDENCIES:'
-' Stage-3 is a root stage'
-' Stage-8 depends on stages: Stage-3 , consists of Stage-5, Stage-4, Stage-6'
-' Stage-5'
-' Stage-2 depends on stages: Stage-5, Stage-4, Stage-7'
-' Stage-1 depends on stages: Stage-2'
-' Stage-4'
-' Stage-6'
-' Stage-7 depends on stages: Stage-6'
-' Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-' Stage: Stage-3'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' index_compression__src_src_index__ '
-' TableScan'
-' alias: index_compression__src_src_index__'
-' filterExpr:'
-' expr: ((key > 80.0) and (key < 100.0))'
-' type: boolean'
-' Filter Operator'
-' predicate:'
-' expr: ((key > 80.0) and (key < 100.0))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: _bucketname'
-' type: string'
-' expr: _offsets'
-' type: array<bigint>'
-' outputColumnNames: _col0, _col1'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 1'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-8'
-' Conditional Operator'
-''
-' Stage: Stage-5'
-' Move Operator'
-' files:'
-' hdfs directory: true'
-' destination: file:!!{hive.exec.scratchdir}!!'
-''
-' Stage: Stage-2'
-' Move Operator'
-' files:'
-' hdfs directory: true'
-' destination: file:!!{hive.exec.scratchdir}!!'
-''
-' Stage: Stage-1'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' src '
-' TableScan'
-' alias: src'
-' filterExpr:'
-' expr: ((key > 80.0) and (key < 100.0))'
-' type: boolean'
-' Filter Operator'
-' predicate:'
-' expr: ((key > 80.0) and (key < 100.0))'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Reduce Output Operator'
-' key expressions:'
-' expr: _col0'
-' type: string'
-' sort order: +'
-' tag: -1'
-' value expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: string'
-' Reduce Operator Tree:'
-' Extract'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-4'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' file:!!{hive.exec.scratchdir}!! '
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-6'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' file:!!{hive.exec.scratchdir}!! '
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-7'
-' Move Operator'
-' files:'
-' hdfs directory: true'
-' destination: file:!!{hive.exec.scratchdir}!!'
-''
-' Stage: Stage-0'
-' Fetch Operator'
-' limit: -1'
-''
-''
-130 rows selected
->>> SELECT key, value FROM src WHERE key > 80 AND key < 100 ORDER BY key;
-'key','value'
-'82','val_82'
-'83','val_83'
-'83','val_83'
-'84','val_84'
-'84','val_84'
-'85','val_85'
-'86','val_86'
-'87','val_87'
-'90','val_90'
-'90','val_90'
-'90','val_90'
-'92','val_92'
-'95','val_95'
-'95','val_95'
-'96','val_96'
-'97','val_97'
-'97','val_97'
-'98','val_98'
-'98','val_98'
-19 rows selected
->>>
->>> DROP INDEX src_index on src;
-No rows affected
->>> !record
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/index_stale.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/index_stale.q.out b/ql/src/test/results/beelinepositive/index_stale.q.out
deleted file mode 100644
index 708b504..0000000
--- a/ql/src/test/results/beelinepositive/index_stale.q.out
+++ /dev/null
@@ -1,79 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/index_stale.q.raw". Enter "record" with no arguments to stop it.
->>> !run !!{qFileDirectory}!!/index_stale.q
->>> -- test that stale indexes are not used
->>>
->>> CREATE TABLE temp(key STRING, val STRING) STORED AS TEXTFILE;
-No rows affected
->>> INSERT OVERWRITE TABLE temp SELECT * FROM src WHERE key < 50;
-'key','value'
-No rows selected
->>>
->>> -- Build an index on temp.
->>> CREATE INDEX temp_index ON TABLE temp(key) as 'COMPACT' WITH DEFERRED REBUILD;
-No rows affected
->>> ALTER INDEX temp_index ON temp REBUILD;
-No rows affected
->>>
->>> SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
-No rows affected
->>> SET hive.optimize.index.filter=true;
-No rows affected
->>> SET hive.optimize.index.filter.compact.minsize=0;
-No rows affected
->>>
->>> -- overwrite temp table so index is out of date
->>> INSERT OVERWRITE TABLE temp SELECT * FROM src;
-'key','value'
-No rows selected
->>>
->>> -- should return correct results bypassing index
->>> EXPLAIN SELECT * FROM temp WHERE key = 86;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME temp))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (= (TOK_TABLE_OR_COL key) 86))))'
-''
-'STAGE DEPENDENCIES:'
-' Stage-1 is a root stage'
-' Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-' Stage: Stage-1'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' temp '
-' TableScan'
-' alias: temp'
-' filterExpr:'
-' expr: (key = 86.0)'
-' type: boolean'
-' Filter Operator'
-' predicate:'
-' expr: (key = 86.0)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: val'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-0'
-' Fetch Operator'
-' limit: -1'
-''
-''
-40 rows selected
->>> SELECT * FROM temp WHERE key = 86;
-'key','val'
-'86','val_86'
-1 row selected
->>> DROP table temp;
-No rows affected
->>> !record