You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by gu...@apache.org on 2017/02/03 21:50:37 UTC
[24/51] [partial] hive git commit: HIVE-15790: Remove unused beeline
golden files (Gunther Hagleitner, reviewed by Sergey Shelukhin)
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/create_like_view.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/create_like_view.q.out b/ql/src/test/results/beelinepositive/create_like_view.q.out
deleted file mode 100644
index 4d5ede1..0000000
--- a/ql/src/test/results/beelinepositive/create_like_view.q.out
+++ /dev/null
@@ -1,203 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/create_like_view.q.raw". Enter "record" with no arguments to stop it.
->>> !run !!{qFileDirectory}!!/create_like_view.q
->>> DROP TABLE IF EXISTS table1;
-No rows affected
->>> DROP TABLE IF EXISTS table2;
-No rows affected
->>> DROP TABLE IF EXISTS table3;
-No rows affected
->>> DROP VIEW IF EXISTS view1;
-No rows affected
->>>
->>> CREATE TABLE table1 (a STRING, b STRING) STORED AS TEXTFILE;
-No rows affected
->>> DESCRIBE table1;
-'col_name','data_type','comment'
-'a','string',''
-'b','string',''
-2 rows selected
->>> DESCRIBE FORMATTED table1;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'a ','string ','None '
-'b ','string ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_like_view ',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Location: ','!!{hive.metastore.warehouse.dir}!!/create_like_view.db/table1',''
-'Table Type: ','MANAGED_TABLE ',''
-'Table Parameters:','',''
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe',''
-'InputFormat: ','org.apache.hadoop.mapred.TextInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'Storage Desc Params:','',''
-'','serialization.format','1 '
-27 rows selected
->>>
->>> CREATE VIEW view1 AS SELECT * FROM table1;
-'a','b'
-No rows selected
->>>
->>> CREATE TABLE table2 LIKE view1;
-No rows affected
->>> DESCRIBE table2;
-'col_name','data_type','comment'
-'a','string',''
-'b','string',''
-2 rows selected
->>> DESCRIBE FORMATTED table2;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'a ','string ','None '
-'b ','string ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_like_view ',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Location: ','!!{hive.metastore.warehouse.dir}!!/create_like_view.db/table2',''
-'Table Type: ','MANAGED_TABLE ',''
-'Table Parameters:','',''
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe',''
-'InputFormat: ','org.apache.hadoop.mapred.TextInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'Storage Desc Params:','',''
-'','serialization.format','1 '
-27 rows selected
->>>
->>> CREATE TABLE IF NOT EXISTS table2 LIKE view1;
-No rows affected
->>>
->>> CREATE EXTERNAL TABLE IF NOT EXISTS table2 LIKE view1;
-No rows affected
->>>
->>> CREATE EXTERNAL TABLE IF NOT EXISTS table3 LIKE view1;
-No rows affected
->>> DESCRIBE table3;
-'col_name','data_type','comment'
-'a','string',''
-'b','string',''
-2 rows selected
->>> DESCRIBE FORMATTED table3;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'a ','string ','None '
-'b ','string ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_like_view ',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Location: ','!!{hive.metastore.warehouse.dir}!!/create_like_view.db/table3',''
-'Table Type: ','EXTERNAL_TABLE ',''
-'Table Parameters:','',''
-'','EXTERNAL ','TRUE '
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe',''
-'InputFormat: ','org.apache.hadoop.mapred.TextInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'Storage Desc Params:','',''
-'','serialization.format','1 '
-28 rows selected
->>>
->>> INSERT OVERWRITE TABLE table1 SELECT key, value FROM src WHERE key = 86;
-'key','value'
-No rows selected
->>> INSERT OVERWRITE TABLE table2 SELECT key, value FROM src WHERE key = 100;
-'key','value'
-No rows selected
->>>
->>> SELECT * FROM table1 order by a, b;
-'a','b'
-'86','val_86'
-1 row selected
->>> SELECT * FROM table2 order by a, b;
-'a','b'
-'100','val_100'
-'100','val_100'
-2 rows selected
->>>
->>> DROP TABLE table1;
-No rows affected
->>> DROP TABLE table2;
-No rows affected
->>> DROP VIEW view1;
-No rows affected
->>>
->>> -- check partitions
->>> create view view1 partitioned on (ds, hr) as select * from srcpart;
-'key','value','ds','hr'
-No rows selected
->>> create table table1 like view1;
-No rows affected
->>> describe formatted table1;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'key ','string ','None '
-'value ','string ','None '
-'','',''
-'# Partition Information','',''
-'# col_name ','data_type ','comment '
-'','',''
-'ds ','string ','None '
-'hr ','string ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_like_view ',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Location: ','!!{hive.metastore.warehouse.dir}!!/create_like_view.db/table1',''
-'Table Type: ','MANAGED_TABLE ',''
-'Table Parameters:','',''
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe',''
-'InputFormat: ','org.apache.hadoop.mapred.TextInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'Storage Desc Params:','',''
-'','serialization.format','1 '
-33 rows selected
->>> DROP TABLE table1;
-No rows affected
->>> DROP VIEW view1;
-No rows affected
->>> !record
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/create_merge_compressed.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/create_merge_compressed.q.out b/ql/src/test/results/beelinepositive/create_merge_compressed.q.out
deleted file mode 100644
index 43e648e..0000000
--- a/ql/src/test/results/beelinepositive/create_merge_compressed.q.out
+++ /dev/null
@@ -1,84 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/create_merge_compressed.q.raw". Enter "record" with no arguments to stop it.
->>> !run !!{qFileDirectory}!!/create_merge_compressed.q
->>> create table src_rc_merge_test(key int, value string) stored as rcfile;
-No rows affected
->>>
->>> load data local inpath '../data/files/smbbucket_1.rc' into table src_rc_merge_test;
-No rows affected
->>>
->>> set hive.exec.compress.output = true;
-No rows affected
->>>
->>> create table tgt_rc_merge_test(key int, value string) stored as rcfile;
-No rows affected
->>> insert into table tgt_rc_merge_test select * from src_rc_merge_test;
-'key','value'
-No rows selected
->>> insert into table tgt_rc_merge_test select * from src_rc_merge_test;
-'key','value'
-No rows selected
->>>
->>> show table extended like `tgt_rc_merge_test`;
-'tab_name'
-'tableName:tgt_rc_merge_test'
-'owner:!!{user.name}!!'
-'location:!!{hive.metastore.warehouse.dir}!!/create_merge_compressed.db/tgt_rc_merge_test'
-'inputformat:org.apache.hadoop.hive.ql.io.RCFileInputFormat'
-'outputformat:org.apache.hadoop.hive.ql.io.RCFileOutputFormat'
-'columns:struct columns { i32 key, string value}'
-'partitioned:false'
-'partitionColumns:'
-'totalNumberFiles:1'
-'totalFileSize:171'
-'maxFileSize:171'
-'minFileSize:171'
-'lastAccessTime:0'
-'lastUpdateTime:!!UNIXTIMEMILLIS!!'
-''
-15 rows selected
->>>
->>> select count(1) from tgt_rc_merge_test;
-'_c0'
-'5'
-1 row selected
->>> select sum(hash(key)), sum(hash(value)) from tgt_rc_merge_test;
-'_c0','_c1'
-'23','-375947694'
-1 row selected
->>>
->>> alter table tgt_rc_merge_test concatenate;
-No rows affected
->>>
->>> show table extended like `tgt_rc_merge_test`;
-'tab_name'
-'tableName:tgt_rc_merge_test'
-'owner:!!{user.name}!!'
-'location:!!{hive.metastore.warehouse.dir}!!/create_merge_compressed.db/tgt_rc_merge_test'
-'inputformat:org.apache.hadoop.hive.ql.io.RCFileInputFormat'
-'outputformat:org.apache.hadoop.hive.ql.io.RCFileOutputFormat'
-'columns:struct columns { i32 key, string value}'
-'partitioned:false'
-'partitionColumns:'
-'totalNumberFiles:1'
-'totalFileSize:171'
-'maxFileSize:171'
-'minFileSize:171'
-'lastAccessTime:0'
-'lastUpdateTime:!!UNIXTIMEMILLIS!!'
-''
-15 rows selected
->>>
->>> select count(1) from tgt_rc_merge_test;
-'_c0'
-'5'
-1 row selected
->>> select sum(hash(key)), sum(hash(value)) from tgt_rc_merge_test;
-'_c0','_c1'
-'23','-375947694'
-1 row selected
->>>
->>> drop table src_rc_merge_test;
-No rows affected
->>> drop table tgt_rc_merge_test;
-No rows affected
->>> !record
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/create_skewed_table1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/create_skewed_table1.q.out b/ql/src/test/results/beelinepositive/create_skewed_table1.q.out
deleted file mode 100644
index c887e28..0000000
--- a/ql/src/test/results/beelinepositive/create_skewed_table1.q.out
+++ /dev/null
@@ -1,111 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/create_skewed_table1.q.raw". Enter "record" with no arguments to stop it.
->>> !run !!{qFileDirectory}!!/create_skewed_table1.q
->>> set hive.internal.ddl.list.bucketing.enable=true;
-No rows affected
->>> CREATE TABLE list_bucket_single (key STRING, value STRING) SKEWED BY (key) ON ('1','5','6');
-No rows affected
->>> CREATE TABLE list_bucket_single_2 (key STRING, value STRING) SKEWED BY (key) ON ((1),(5),(6));
-No rows affected
->>> CREATE TABLE list_bucket_multiple (col1 STRING, col2 int, col3 STRING) SKEWED BY (col1, col2) ON (('s1',1), ('s3',3), ('s13',13), ('s78',78));
-No rows affected
->>> describe formatted list_bucket_single_2;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'key ','string ','None '
-'value ','string ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_skewed_table1',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Location: ','!!{hive.metastore.warehouse.dir}!!/create_skewed_table1.db/list_bucket_single_2',''
-'Table Type: ','MANAGED_TABLE ',''
-'Table Parameters:','',''
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe',''
-'InputFormat: ','org.apache.hadoop.mapred.TextInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'Skewed Columns: ','[key] ',''
-'Skewed Values: ','[[1], [5], [6]] ',''
-'Storage Desc Params:','',''
-'','serialization.format','1 '
-29 rows selected
->>> describe formatted list_bucket_single;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'key ','string ','None '
-'value ','string ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_skewed_table1',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Location: ','!!{hive.metastore.warehouse.dir}!!/create_skewed_table1.db/list_bucket_single',''
-'Table Type: ','MANAGED_TABLE ',''
-'Table Parameters:','',''
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe',''
-'InputFormat: ','org.apache.hadoop.mapred.TextInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'Skewed Columns: ','[key] ',''
-'Skewed Values: ','[[1], [5], [6]] ',''
-'Storage Desc Params:','',''
-'','serialization.format','1 '
-29 rows selected
->>> describe formatted list_bucket_multiple;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'col1 ','string ','None '
-'col2 ','int ','None '
-'col3 ','string ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_skewed_table1',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Location: ','!!{hive.metastore.warehouse.dir}!!/create_skewed_table1.db/list_bucket_multiple',''
-'Table Type: ','MANAGED_TABLE ',''
-'Table Parameters:','',''
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe',''
-'InputFormat: ','org.apache.hadoop.mapred.TextInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'Skewed Columns: ','[col1, col2] ',''
-'Skewed Values: ','[[s1, 1], [s3, 3], [s13, 13], [s78, 78]]',''
-'Storage Desc Params:','',''
-'','serialization.format','1 '
-30 rows selected
->>> drop table list_bucket_single;
-No rows affected
->>> drop table list_bucket_multiple;
-No rows affected
->>> drop table list_bucket_single_2;
-No rows affected
->>> !record
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/create_udaf.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/create_udaf.q.out b/ql/src/test/results/beelinepositive/create_udaf.q.out
deleted file mode 100644
index f73919c..0000000
--- a/ql/src/test/results/beelinepositive/create_udaf.q.out
+++ /dev/null
@@ -1,35 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/create_udaf.q.raw". Enter "record" with no arguments to stop it.
->>> !run !!{qFileDirectory}!!/create_udaf.q
->>> EXPLAIN
-CREATE TEMPORARY FUNCTION test_max AS 'org.apache.hadoop.hive.ql.udf.UDAFTestMax';
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_CREATEFUNCTION test_max 'org.apache.hadoop.hive.ql.udf.UDAFTestMax')'
-''
-'STAGE DEPENDENCIES:'
-' Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-' Stage: Stage-0'
-''
-''
-10 rows selected
->>>
->>> CREATE TEMPORARY FUNCTION test_max AS 'org.apache.hadoop.hive.ql.udf.UDAFTestMax';
-No rows affected
->>>
->>> CREATE TABLE dest1(col INT);
-No rows affected
->>>
->>> FROM src INSERT OVERWRITE TABLE dest1 SELECT test_max(length(src.value));
-'_c0'
-No rows selected
->>>
->>> SELECT dest1.* FROM dest1;
-'col'
-'7'
-1 row selected
->>>
->>> DROP TEMPORARY FUNCTION test_max;
-No rows affected
->>> !record
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/create_view.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/create_view.q.out b/ql/src/test/results/beelinepositive/create_view.q.out
deleted file mode 100644
index 2ae4e08..0000000
--- a/ql/src/test/results/beelinepositive/create_view.q.out
+++ /dev/null
@@ -1,1164 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/create_view.q.raw". Enter "record" with no arguments to stop it.
->>> !run !!{qFileDirectory}!!/create_view.q
->>> DROP VIEW view1;
-No rows affected
->>> DROP VIEW view2;
-No rows affected
->>> DROP VIEW view3;
-No rows affected
->>> DROP VIEW view4;
-No rows affected
->>> DROP VIEW view5;
-No rows affected
->>> DROP VIEW view6;
-No rows affected
->>> DROP VIEW view7;
-No rows affected
->>> DROP VIEW view8;
-No rows affected
->>> DROP VIEW view9;
-No rows affected
->>> DROP VIEW view10;
-No rows affected
->>> DROP VIEW view11;
-No rows affected
->>> DROP VIEW view12;
-No rows affected
->>> DROP VIEW view13;
-No rows affected
->>> DROP VIEW view14;
-No rows affected
->>> DROP VIEW view15;
-No rows affected
->>> DROP VIEW view16;
-No rows affected
->>> DROP TEMPORARY FUNCTION test_translate;
-No rows affected
->>> DROP TEMPORARY FUNCTION test_max;
-No rows affected
->>> DROP TEMPORARY FUNCTION test_explode;
-No rows affected
->>>
->>>
->>> SELECT * FROM src WHERE key=86;
-'key','value'
-'86','val_86'
-1 row selected
->>> CREATE VIEW view1 AS SELECT value FROM src WHERE key=86;
-'value'
-No rows selected
->>> CREATE VIEW view2 AS SELECT * FROM src;
-'key','value'
-No rows selected
->>> CREATE VIEW view3(valoo)
-TBLPROPERTIES ("fear" = "factor")
-AS SELECT upper(value) FROM src WHERE key=86;
-'_c0'
-No rows selected
->>> SELECT * from view1;
-'value'
-'val_86'
-1 row selected
->>> SELECT * from view2 where key=18;
-'key','value'
-'18','val_18'
-'18','val_18'
-2 rows selected
->>> SELECT * from view3;
-'valoo'
-'VAL_86'
-1 row selected
->>>
->>> -- test EXPLAIN output for CREATE VIEW
->>> EXPLAIN
-CREATE VIEW view0(valoo) AS SELECT upper(value) FROM src WHERE key=86;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_CREATEVIEW (TOK_TABNAME view0) (TOK_TABCOLNAME (TOK_TABCOL valoo TOK_NULL)) (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION upper (TOK_TABLE_OR_COL value)))) (TOK_WHERE (= (TOK_TABLE_OR_COL key) 86)))))'
-''
-'STAGE DEPENDENCIES:'
-' Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-' Stage: Stage-0'
-' Create View Operator:'
-' Create View'
-' if not exists: false'
-' or replace: false'
-' columns: valoo string'
-' expanded text: SELECT `_c0` AS `valoo` FROM (SELECT upper(`src`.`value`) FROM `create_view`.`src` WHERE `src`.`key`=86) `view0`'
-' name: view0'
-' original text: SELECT upper(value) FROM src WHERE key=86'
-''
-''
-18 rows selected
->>>
->>> -- make sure EXPLAIN works with a query which references a view
->>> EXPLAIN
-SELECT * from view2 where key=18;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-' (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME view2))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (= (TOK_TABLE_OR_COL key) 18))))'
-''
-'STAGE DEPENDENCIES:'
-' Stage-1 is a root stage'
-' Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-' Stage: Stage-1'
-' Map Reduce'
-' Alias -> Map Operator Tree:'
-' view2:src '
-' TableScan'
-' alias: src'
-' Filter Operator'
-' predicate:'
-' expr: (key = 18)'
-' type: boolean'
-' Select Operator'
-' expressions:'
-' expr: key'
-' type: string'
-' expr: value'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' Select Operator'
-' expressions:'
-' expr: _col0'
-' type: string'
-' expr: _col1'
-' type: string'
-' outputColumnNames: _col0, _col1'
-' File Output Operator'
-' compressed: false'
-' GlobalTableId: 0'
-' table:'
-' input format: org.apache.hadoop.mapred.TextInputFormat'
-' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-' Stage: Stage-0'
-' Fetch Operator'
-' limit: -1'
-''
-''
-44 rows selected
->>>
->>> SHOW TABLES 'view.*';
-'tab_name'
-'view1'
-'view2'
-'view3'
-3 rows selected
->>> DESCRIBE view1;
-'col_name','data_type','comment'
-'value','string',''
-1 row selected
->>> DESCRIBE EXTENDED view1;
-'col_name','data_type','comment'
-'value','string',''
-'','',''
-'Detailed Table Information','Table(tableName:view1, dbName:create_view, owner:!!{user.name}!!, createTime:!!UNIXTIME!!, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:value, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[], parameters:{transient_lastDdlTime=!!UNIXTIME!!}, viewOriginalText:SELECT value FROM src WHERE key=86, viewExpandedText:SELECT `src`.`value` FROM `create_view`.`src` WHERE `src`.`key`=86, tableType:VIRTUAL_VIEW)',''
-3 rows selected
->>> DESCRIBE FORMATTED view1;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'value ','string ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_view ',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Table Type: ','VIRTUAL_VIEW ',''
-'Table Parameters:','',''
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','null ',''
-'InputFormat: ','org.apache.hadoop.mapred.SequenceFileInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'','',''
-'# View Information','',''
-'View Original Text: ','SELECT value FROM src WHERE key=86',''
-'View Expanded Text: ','SELECT `src`.`value` FROM `create_view`.`src` WHERE `src`.`key`=86',''
-27 rows selected
->>> DESCRIBE view2;
-'col_name','data_type','comment'
-'key','string',''
-'value','string',''
-2 rows selected
->>> DESCRIBE EXTENDED view2;
-'col_name','data_type','comment'
-'key','string',''
-'value','string',''
-'','',''
-'Detailed Table Information','Table(tableName:view2, dbName:create_view, owner:!!{user.name}!!, createTime:!!UNIXTIME!!, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[], parameters:{transient_lastDdlTime=!!UNIXTIME!!}, viewOriginalText:SELECT * FROM src, viewExpandedText:SELECT `src`.`key`, `src`.`value` FROM `create_view`.`src`, tableType:VIRTUAL_VIEW)',''
-4 rows selected
->>> DESCRIBE FORMATTED view2;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'key ','string ','None '
-'value ','string ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_view ',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Table Type: ','VIRTUAL_VIEW ',''
-'Table Parameters:','',''
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','null ',''
-'InputFormat: ','org.apache.hadoop.mapred.SequenceFileInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'','',''
-'# View Information','',''
-'View Original Text: ','SELECT * FROM src ',''
-'View Expanded Text: ','SELECT `src`.`key`, `src`.`value` FROM `create_view`.`src`',''
-28 rows selected
->>> DESCRIBE view3;
-'col_name','data_type','comment'
-'valoo','string',''
-1 row selected
->>> DESCRIBE EXTENDED view3;
-'col_name','data_type','comment'
-'valoo','string',''
-'','',''
-'Detailed Table Information','Table(tableName:view3, dbName:create_view, owner:!!{user.name}!!, createTime:!!UNIXTIME!!, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:valoo, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[], parameters:{transient_lastDdlTime=!!UNIXTIME!!, fear=factor}, viewOriginalText:SELECT upper(value) FROM src WHERE key=86, viewExpandedText:SELECT `_c0` AS `valoo` FROM (SELECT upper(`src`.`value`) FROM `create_view`.`src` WHERE `src`.`key`=86) `view3`, tableType:VIRTUAL_VIEW)',''
-3 rows selected
->>> DESCRIBE FORMATTED view3;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'valoo ','string ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_view ',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Table Type: ','VIRTUAL_VIEW ',''
-'Table Parameters:','',''
-'','fear ','factor '
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','null ',''
-'InputFormat: ','org.apache.hadoop.mapred.SequenceFileInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'','',''
-'# View Information','',''
-'View Original Text: ','SELECT upper(value) FROM src WHERE key=86',''
-'View Expanded Text: ','SELECT `_c0` AS `valoo` FROM (SELECT upper(`src`.`value`) FROM `create_view`.`src` WHERE `src`.`key`=86) `view3`',''
-28 rows selected
->>>
->>> ALTER VIEW view3 SET TBLPROPERTIES ("biggest" = "loser");
-No rows affected
->>> DESCRIBE EXTENDED view3;
-'col_name','data_type','comment'
-'valoo','string',''
-'','',''
-'Detailed Table Information','Table(tableName:view3, dbName:create_view, owner:!!{user.name}!!, createTime:!!UNIXTIME!!, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:valoo, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[], parameters:{last_modified_by=!!ELIDED!!, last_modified_time=!!UNIXTIME!!, biggest=loser, transient_lastDdlTime=!!UNIXTIME!!, fear=factor}, viewOriginalText:SELECT upper(value) FROM src WHERE key=86, viewExpandedText:SELECT `_c0` AS `valoo` FROM (SELECT upper(`src`.`value`) FROM `create_view`.`src` WHERE `src`.`key`=86) `view3`, tableType:VIRTUAL_VIEW)',''
-3 rows selected
->>> DESCRIBE FORMATTED view3;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'valoo ','string ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_view ',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Table Type: ','VIRTUAL_VIEW ',''
-'Table Parameters:','',''
-'','biggest ','loser '
-'','fear ','factor '
-'','last_modified_by ','!!{user.name}!! '
-'','last_modified_time ','!!UNIXTIME!! '
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','null ',''
-'InputFormat: ','org.apache.hadoop.mapred.SequenceFileInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'','',''
-'# View Information','',''
-'View Original Text: ','SELECT upper(value) FROM src WHERE key=86',''
-'View Expanded Text: ','SELECT `_c0` AS `valoo` FROM (SELECT upper(`src`.`value`) FROM `create_view`.`src` WHERE `src`.`key`=86) `view3`',''
-31 rows selected
->>>
->>> CREATE TABLE table1 (key int);
-No rows affected
->>>
->>> -- use DESCRIBE EXTENDED on a base table and an external table as points
->>> -- of comparison for view descriptions
->>> DESCRIBE EXTENDED table1;
-'col_name','data_type','comment'
-'key','int',''
-'','',''
-'Detailed Table Information','Table(tableName:table1, dbName:create_view, owner:!!{user.name}!!, createTime:!!UNIXTIME!!, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:int, comment:null)], location:!!{hive.metastore.warehouse.dir}!!/create_view.db/table1, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[], parameters:{transient_lastDdlTime=!!UNIXTIME!!}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)',''
-3 rows selected
->>> DESCRIBE EXTENDED src1;
-'col_name','data_type','comment'
-'key','string',''
-'value','string',''
-'','',''
-'Detailed Table Information','Table(tableName:src1, dbName:create_view, owner:!!{user.name}!!, createTime:!!UNIXTIME!!, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:!!{hive.metastore.warehouse.dir}!!/create_view.db/src1, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[], parameters:{numPartitions=0, numFiles=1, transient_lastDdlTime=!!UNIXTIME!!, totalSize=216, numRows=0, rawDataSize=0}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)',''
-4 rows selected
->>>
->>> -- use DESCRIBE EXTENDED on a base table as a point of comparison for
->>> -- view descriptions
->>> DESCRIBE EXTENDED table1;
-'col_name','data_type','comment'
-'key','int',''
-'','',''
-'Detailed Table Information','Table(tableName:table1, dbName:create_view, owner:!!{user.name}!!, createTime:!!UNIXTIME!!, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:int, comment:null)], location:!!{hive.metastore.warehouse.dir}!!/create_view.db/table1, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[], parameters:{transient_lastDdlTime=!!UNIXTIME!!}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)',''
-3 rows selected
->>>
->>>
->>> INSERT OVERWRITE TABLE table1 SELECT key FROM src WHERE key = 86;
-'_col0'
-No rows selected
->>>
->>> SELECT * FROM table1;
-'key'
-'86'
-1 row selected
->>> CREATE VIEW view4 AS SELECT * FROM table1;
-'key'
-No rows selected
->>> SELECT * FROM view4;
-'key'
-'86'
-1 row selected
->>> DESCRIBE view4;
-'col_name','data_type','comment'
-'key','int',''
-1 row selected
->>> ALTER TABLE table1 ADD COLUMNS (value STRING);
-No rows affected
->>> SELECT * FROM table1;
-'key','value'
-'86',''
-1 row selected
->>> SELECT * FROM view4;
-'key'
-'86'
-1 row selected
->>> DESCRIBE table1;
-'col_name','data_type','comment'
-'key','int',''
-'value','string',''
-2 rows selected
->>> DESCRIBE view4;
-'col_name','data_type','comment'
-'key','int',''
-1 row selected
->>>
->>> CREATE VIEW view5 AS SELECT v1.key as key1, v2.key as key2
-FROM view4 v1 join view4 v2;
-'key1','key2'
-No rows selected
->>> SELECT * FROM view5;
-'key1','key2'
-'86','86'
-1 row selected
->>> DESCRIBE view5;
-'col_name','data_type','comment'
-'key1','int',''
-'key2','int',''
-2 rows selected
->>>
->>> -- verify that column name and comment in DDL portion
->>> -- overrides column alias in SELECT
->>> CREATE VIEW view6(valoo COMMENT 'I cannot spell') AS
-SELECT upper(value) as blarg FROM src WHERE key=86;
-'blarg'
-No rows selected
->>> DESCRIBE view6;
-'col_name','data_type','comment'
-'valoo','string','I cannot spell'
-1 row selected
->>>
->>> -- verify that ORDER BY and LIMIT are both supported in view def
->>> CREATE VIEW view7 AS
-SELECT * FROM src
-WHERE key > 80 AND key < 100
-ORDER BY key, value
-LIMIT 10;
-'key','value'
-No rows selected
->>>
->>> SELECT * FROM view7;
-'key','value'
-'82','val_82'
-'83','val_83'
-'83','val_83'
-'84','val_84'
-'84','val_84'
-'85','val_85'
-'86','val_86'
-'87','val_87'
-'90','val_90'
-'90','val_90'
-10 rows selected
->>>
->>> -- top-level ORDER BY should override the one inside the view
->>> -- (however, the inside ORDER BY should still influence the evaluation
->>> -- of the limit)
->>> SELECT * FROM view7 ORDER BY key DESC, value;
-'key','value'
-'90','val_90'
-'90','val_90'
-'87','val_87'
-'86','val_86'
-'85','val_85'
-'84','val_84'
-'84','val_84'
-'83','val_83'
-'83','val_83'
-'82','val_82'
-10 rows selected
->>>
->>> -- top-level LIMIT should override if lower
->>> SELECT * FROM view7 LIMIT 5;
-'key','value'
-'82','val_82'
-'83','val_83'
-'83','val_83'
-'84','val_84'
-'84','val_84'
-5 rows selected
->>>
->>> -- but not if higher
->>> SELECT * FROM view7 LIMIT 20;
-'key','value'
-'82','val_82'
-'83','val_83'
-'83','val_83'
-'84','val_84'
-'84','val_84'
-'85','val_85'
-'86','val_86'
-'87','val_87'
-'90','val_90'
-'90','val_90'
-10 rows selected
->>>
->>> -- test usage of a function within a view
->>> CREATE TEMPORARY FUNCTION test_translate AS
-'org.apache.hadoop.hive.ql.udf.generic.GenericUDFTestTranslate';
-No rows affected
->>> CREATE VIEW view8(c) AS
-SELECT test_translate('abc', 'a', 'b')
-FROM table1;
-'_c0'
-No rows selected
->>> DESCRIBE EXTENDED view8;
-'col_name','data_type','comment'
-'c','string',''
-'','',''
-'Detailed Table Information','Table(tableName:view8, dbName:create_view, owner:!!{user.name}!!, createTime:!!UNIXTIME!!, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:c, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[], parameters:{transient_lastDdlTime=!!UNIXTIME!!}, viewOriginalText:SELECT test_translate('abc', 'a', 'b') ',''
-'FROM table1, viewExpandedText:SELECT `_c0` AS `c` FROM (SELECT `test_translate`('abc', 'a', 'b') ','',''
-'FROM `create_view`.`table1`) `view8`, tableType:VIRTUAL_VIEW)','',''
-5 rows selected
->>> DESCRIBE FORMATTED view8;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'c ','string ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_view ',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Table Type: ','VIRTUAL_VIEW ',''
-'Table Parameters:','',''
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','null ',''
-'InputFormat: ','org.apache.hadoop.mapred.SequenceFileInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'','',''
-'# View Information','',''
-'View Original Text: ','SELECT test_translate('abc', 'a', 'b') ',''
-'FROM table1','',''
-'View Expanded Text: ','SELECT `_c0` AS `c` FROM (SELECT `test_translate`('abc', 'a', 'b') ',''
-'FROM `create_view`.`table1`) `view8`','',''
-29 rows selected
->>> SELECT * FROM view8;
-'c'
-'bbc'
-1 row selected
->>>
->>> -- test usage of a UDAF within a view
->>> CREATE TEMPORARY FUNCTION test_max AS
-'org.apache.hadoop.hive.ql.udf.UDAFTestMax';
-No rows affected
->>> set hive.map.aggr=false;
-No rows affected
->>> -- disable map-side aggregation
->>> CREATE VIEW view9(m) AS
-SELECT test_max(length(value))
-FROM src;
-'_c0'
-No rows selected
->>> DESCRIBE EXTENDED view9;
-'col_name','data_type','comment'
-'m','int',''
-'','',''
-'Detailed Table Information','Table(tableName:view9, dbName:create_view, owner:!!{user.name}!!, createTime:!!UNIXTIME!!, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:m, type:int, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[], parameters:{transient_lastDdlTime=!!UNIXTIME!!}, viewOriginalText:SELECT test_max(length(value)) ',''
-'FROM src, viewExpandedText:SELECT `_c0` AS `m` FROM (SELECT `test_max`(length(`src`.`value`)) ','',''
-'FROM `create_view`.`src`) `view9`, tableType:VIRTUAL_VIEW)','',''
-5 rows selected
->>> DESCRIBE FORMATTED view9;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'m ','int ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_view ',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Table Type: ','VIRTUAL_VIEW ',''
-'Table Parameters:','',''
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','null ',''
-'InputFormat: ','org.apache.hadoop.mapred.SequenceFileInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'','',''
-'# View Information','',''
-'View Original Text: ','SELECT test_max(length(value)) ',''
-'FROM src','',''
-'View Expanded Text: ','SELECT `_c0` AS `m` FROM (SELECT `test_max`(length(`src`.`value`)) ',''
-'FROM `create_view`.`src`) `view9`','',''
-29 rows selected
->>> SELECT * FROM view9;
-'m'
-'7'
-1 row selected
->>> DROP VIEW view9;
-No rows affected
->>> set hive.map.aggr=true;
-No rows affected
->>> -- enable map-side aggregation
->>> CREATE VIEW view9(m) AS
-SELECT test_max(length(value))
-FROM src;
-'_c0'
-No rows selected
->>> DESCRIBE EXTENDED view9;
-'col_name','data_type','comment'
-'m','int',''
-'','',''
-'Detailed Table Information','Table(tableName:view9, dbName:create_view, owner:!!{user.name}!!, createTime:!!UNIXTIME!!, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:m, type:int, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[], parameters:{transient_lastDdlTime=!!UNIXTIME!!}, viewOriginalText:SELECT test_max(length(value)) ',''
-'FROM src, viewExpandedText:SELECT `_c0` AS `m` FROM (SELECT `test_max`(length(`src`.`value`)) ','',''
-'FROM `create_view`.`src`) `view9`, tableType:VIRTUAL_VIEW)','',''
-5 rows selected
->>> DESCRIBE FORMATTED view9;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'m ','int ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_view ',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Table Type: ','VIRTUAL_VIEW ',''
-'Table Parameters:','',''
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','null ',''
-'InputFormat: ','org.apache.hadoop.mapred.SequenceFileInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'','',''
-'# View Information','',''
-'View Original Text: ','SELECT test_max(length(value)) ',''
-'FROM src','',''
-'View Expanded Text: ','SELECT `_c0` AS `m` FROM (SELECT `test_max`(length(`src`.`value`)) ',''
-'FROM `create_view`.`src`) `view9`','',''
-29 rows selected
->>> SELECT * FROM view9;
-'m'
-'7'
-1 row selected
->>>
->>> -- test usage of a subselect within a view
->>> CREATE VIEW view10 AS
-SELECT slurp.* FROM (SELECT * FROM src WHERE key=86) slurp;
-'key','value'
-No rows selected
->>> DESCRIBE EXTENDED view10;
-'col_name','data_type','comment'
-'key','string',''
-'value','string',''
-'','',''
-'Detailed Table Information','Table(tableName:view10, dbName:create_view, owner:!!{user.name}!!, createTime:!!UNIXTIME!!, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[], parameters:{transient_lastDdlTime=!!UNIXTIME!!}, viewOriginalText:SELECT slurp.* FROM (SELECT * FROM src WHERE key=86) slurp, viewExpandedText:SELECT `slurp`.`key`, `slurp`.`value` FROM (SELECT `src`.`key`, `src`.`value` FROM `create_view`.`src` WHERE `src`.`key`=86) `slurp`, tableType:VIRTUAL_VIEW)',''
-4 rows selected
->>> DESCRIBE FORMATTED view10;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'key ','string ','None '
-'value ','string ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_view ',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Table Type: ','VIRTUAL_VIEW ',''
-'Table Parameters:','',''
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','null ',''
-'InputFormat: ','org.apache.hadoop.mapred.SequenceFileInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'','',''
-'# View Information','',''
-'View Original Text: ','SELECT slurp.* FROM (SELECT * FROM src WHERE key=86) slurp',''
-'View Expanded Text: ','SELECT `slurp`.`key`, `slurp`.`value` FROM (SELECT `src`.`key`, `src`.`value` FROM `create_view`.`src` WHERE `src`.`key`=86) `slurp`',''
-28 rows selected
->>> SELECT * FROM view10;
-'key','value'
-'86','val_86'
-1 row selected
->>>
->>> -- test usage of a UDTF within a view
->>> CREATE TEMPORARY FUNCTION test_explode AS
-'org.apache.hadoop.hive.ql.udf.generic.GenericUDTFExplode';
-No rows affected
->>> CREATE VIEW view11 AS
-SELECT test_explode(array(1,2,3)) AS (boom)
-FROM table1;
-'boom'
-No rows selected
->>> DESCRIBE EXTENDED view11;
-'col_name','data_type','comment'
-'boom','int',''
-'','',''
-'Detailed Table Information','Table(tableName:view11, dbName:create_view, owner:!!{user.name}!!, createTime:!!UNIXTIME!!, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:boom, type:int, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[], parameters:{transient_lastDdlTime=!!UNIXTIME!!}, viewOriginalText:SELECT test_explode(array(1,2,3)) AS (boom) ',''
-'FROM table1, viewExpandedText:SELECT `test_explode`(array(1,2,3)) AS (`boom`) ','',''
-'FROM `create_view`.`table1`, tableType:VIRTUAL_VIEW)','',''
-5 rows selected
->>> DESCRIBE FORMATTED view11;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'boom ','int ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_view ',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Table Type: ','VIRTUAL_VIEW ',''
-'Table Parameters:','',''
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','null ',''
-'InputFormat: ','org.apache.hadoop.mapred.SequenceFileInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'','',''
-'# View Information','',''
-'View Original Text: ','SELECT test_explode(array(1,2,3)) AS (boom) ',''
-'FROM table1','',''
-'View Expanded Text: ','SELECT `test_explode`(array(1,2,3)) AS (`boom`) ',''
-'FROM `create_view`.`table1`','',''
-29 rows selected
->>> SELECT * FROM view11;
-'boom'
-'1'
-'2'
-'3'
-3 rows selected
->>>
->>> -- test usage of LATERAL within a view
->>> CREATE VIEW view12 AS
-SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol;
-'key','value','mycol'
-No rows selected
->>> DESCRIBE EXTENDED view12;
-'col_name','data_type','comment'
-'key','string',''
-'value','string',''
-'mycol','int',''
-'','',''
-'Detailed Table Information','Table(tableName:view12, dbName:create_view, owner:!!{user.name}!!, createTime:!!UNIXTIME!!, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null), FieldSchema(name:mycol, type:int, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[], parameters:{transient_lastDdlTime=!!UNIXTIME!!}, viewOriginalText:SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol, viewExpandedText:SELECT `src`.`key`, `src`.`value`, `mytable`.`mycol` FROM `create_view`.`src` LATERAL VIEW explode(array
(1,2,3)) `myTable` AS `myCol`, tableType:VIRTUAL_VIEW)',''
-5 rows selected
->>> DESCRIBE FORMATTED view12;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'key ','string ','None '
-'value ','string ','None '
-'mycol ','int ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_view ',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Table Type: ','VIRTUAL_VIEW ',''
-'Table Parameters:','',''
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','null ',''
-'InputFormat: ','org.apache.hadoop.mapred.SequenceFileInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'','',''
-'# View Information','',''
-'View Original Text: ','SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol',''
-'View Expanded Text: ','SELECT `src`.`key`, `src`.`value`, `mytable`.`mycol` FROM `create_view`.`src` LATERAL VIEW explode(array(1,2,3)) `myTable` AS `myCol`',''
-29 rows selected
->>> SELECT * FROM view12
-ORDER BY key ASC, myCol ASC LIMIT 1;
-'key','value','mycol'
-'0','val_0','1'
-1 row selected
->>>
->>> -- test usage of LATERAL with a view as the LHS
->>> SELECT * FROM view2 LATERAL VIEW explode(array(1,2,3)) myTable AS myCol
-ORDER BY key ASC, myCol ASC LIMIT 1;
-'key','value','mycol'
-'0','val_0','1'
-1 row selected
->>>
->>> -- test usage of TABLESAMPLE within a view
->>> CREATE VIEW view13 AS
-SELECT s.key
-FROM srcbucket TABLESAMPLE (BUCKET 1 OUT OF 5 ON key) s;
-'key'
-No rows selected
->>> DESCRIBE EXTENDED view13;
-'col_name','data_type','comment'
-'key','int',''
-'','',''
-'Detailed Table Information','Table(tableName:view13, dbName:create_view, owner:!!{user.name}!!, createTime:!!UNIXTIME!!, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:int, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[], parameters:{transient_lastDdlTime=!!UNIXTIME!!}, viewOriginalText:SELECT s.key ',''
-'FROM srcbucket TABLESAMPLE (BUCKET 1 OUT OF 5 ON key) s, viewExpandedText:SELECT `s`.`key` ','',''
-'FROM `create_view`.`srcbucket` TABLESAMPLE (BUCKET 1 OUT OF 5 ON `key`) `s`, tableType:VIRTUAL_VIEW)','',''
-5 rows selected
->>> DESCRIBE FORMATTED view13;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'key ','int ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_view ',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Table Type: ','VIRTUAL_VIEW ',''
-'Table Parameters:','',''
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','null ',''
-'InputFormat: ','org.apache.hadoop.mapred.SequenceFileInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'','',''
-'# View Information','',''
-'View Original Text: ','SELECT s.key ',''
-'FROM srcbucket TABLESAMPLE (BUCKET 1 OUT OF 5 ON key) s','',''
-'View Expanded Text: ','SELECT `s`.`key` ',''
-'FROM `create_view`.`srcbucket` TABLESAMPLE (BUCKET 1 OUT OF 5 ON `key`) `s`','',''
-29 rows selected
->>> SELECT * FROM view13
-ORDER BY key LIMIT 12;
-'key'
-'0'
-'0'
-'0'
-'0'
-'0'
-'5'
-'5'
-'5'
-'5'
-'10'
-'10'
-'15'
-12 rows selected
->>>
->>> -- test usage of JOIN+UNION+AGG all within same view
->>> CREATE VIEW view14 AS
-SELECT unionsrc1.key as k1, unionsrc1.value as v1,
-unionsrc2.key as k2, unionsrc2.value as v2
-FROM (select 'tst1' as key, cast(count(1) as string) as value from src s1
-UNION ALL
-select s2.key as key, s2.value as value from src s2 where s2.key < 10) unionsrc1
-JOIN
-(select 'tst1' as key, cast(count(1) as string) as value from src s3
-UNION ALL
-select s4.key as key, s4.value as value from src s4 where s4.key < 10) unionsrc2
-ON (unionsrc1.key = unionsrc2.key);
-'k1','v1','k2','v2'
-No rows selected
->>> DESCRIBE EXTENDED view14;
-'col_name','data_type','comment'
-'k1','string',''
-'v1','string',''
-'k2','string',''
-'v2','string',''
-'','',''
-'Detailed Table Information','Table(tableName:view14, dbName:create_view, owner:!!{user.name}!!, createTime:!!UNIXTIME!!, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:k1, type:string, comment:null), FieldSchema(name:v1, type:string, comment:null), FieldSchema(name:k2, type:string, comment:null), FieldSchema(name:v2, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[], parameters:{transient_lastDdlTime=!!UNIXTIME!!}, viewOriginalText:SELECT unionsrc1.key as k1, unionsrc1.value as v1, ',''
-'unionsrc2.key as k2, unionsrc2.value as v2 ','',''
-'FROM (select 'tst1' as key, cast(count(1) as string) as value from src s1 ','',''
-'UNION ALL ','',''
-'select s2.key as key, s2.value as value from src s2 where s2.key < 10) unionsrc1 ','',''
-'JOIN ','',''
-'(select 'tst1' as key, cast(count(1) as string) as value from src s3 ','',''
-'UNION ALL ','',''
-'select s4.key as key, s4.value as value from src s4 where s4.key < 10) unionsrc2 ','',''
-'ON (unionsrc1.key = unionsrc2.key), viewExpandedText:SELECT `unionsrc1`.`key` as `k1`, `unionsrc1`.`value` as `v1`, ','',''
-'`unionsrc2`.`key` as `k2`, `unionsrc2`.`value` as `v2` ','',''
-'FROM (select 'tst1' as `key`, cast(count(1) as string) as `value` from `create_view`.`src` `s1` ','',''
-'UNION ALL ','',''
-'select `s2`.`key` as `key`, `s2`.`value` as `value` from `create_view`.`src` `s2` where `s2`.`key` < 10) `unionsrc1` ','',''
-'JOIN ','',''
-'(select 'tst1' as `key`, cast(count(1) as string) as `value` from `create_view`.`src` `s3` ','',''
-'UNION ALL ','',''
-'select `s4`.`key` as `key`, `s4`.`value` as `value` from `create_view`.`src` `s4` where `s4`.`key` < 10) `unionsrc2` ','',''
-'ON (`unionsrc1`.`key` = `unionsrc2`.`key`), tableType:VIRTUAL_VIEW)','',''
-24 rows selected
->>> DESCRIBE FORMATTED view14;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'k1 ','string ','None '
-'v1 ','string ','None '
-'k2 ','string ','None '
-'v2 ','string ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_view ',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Table Type: ','VIRTUAL_VIEW ',''
-'Table Parameters:','',''
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','null ',''
-'InputFormat: ','org.apache.hadoop.mapred.SequenceFileInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'','',''
-'# View Information','',''
-'View Original Text: ','SELECT unionsrc1.key as k1, unionsrc1.value as v1, ',''
-'unionsrc2.key as k2, unionsrc2.value as v2 ','',''
-'FROM (select 'tst1' as key, cast(count(1) as string) as value from src s1 ','',''
-'UNION ALL ','',''
-'select s2.key as key, s2.value as value from src s2 where s2.key < 10) unionsrc1 ','',''
-'JOIN ','',''
-'(select 'tst1' as key, cast(count(1) as string) as value from src s3 ','',''
-'UNION ALL ','',''
-'select s4.key as key, s4.value as value from src s4 where s4.key < 10) unionsrc2 ','',''
-'ON (unionsrc1.key = unionsrc2.key)','',''
-'View Expanded Text: ','SELECT `unionsrc1`.`key` as `k1`, `unionsrc1`.`value` as `v1`, ',''
-'`unionsrc2`.`key` as `k2`, `unionsrc2`.`value` as `v2` ','',''
-'FROM (select 'tst1' as `key`, cast(count(1) as string) as `value` from `create_view`.`src` `s1` ','',''
-'UNION ALL ','',''
-'select `s2`.`key` as `key`, `s2`.`value` as `value` from `create_view`.`src` `s2` where `s2`.`key` < 10) `unionsrc1` ','',''
-'JOIN ','',''
-'(select 'tst1' as `key`, cast(count(1) as string) as `value` from `create_view`.`src` `s3` ','',''
-'UNION ALL ','',''
-'select `s4`.`key` as `key`, `s4`.`value` as `value` from `create_view`.`src` `s4` where `s4`.`key` < 10) `unionsrc2` ','',''
-'ON (`unionsrc1`.`key` = `unionsrc2`.`key`)','',''
-48 rows selected
->>> SELECT * FROM view14
-ORDER BY k1;
-'k1','v1','k2','v2'
-'0','val_0','0','val_0'
-'0','val_0','0','val_0'
-'0','val_0','0','val_0'
-'0','val_0','0','val_0'
-'0','val_0','0','val_0'
-'0','val_0','0','val_0'
-'0','val_0','0','val_0'
-'0','val_0','0','val_0'
-'0','val_0','0','val_0'
-'2','val_2','2','val_2'
-'4','val_4','4','val_4'
-'5','val_5','5','val_5'
-'5','val_5','5','val_5'
-'5','val_5','5','val_5'
-'5','val_5','5','val_5'
-'5','val_5','5','val_5'
-'5','val_5','5','val_5'
-'5','val_5','5','val_5'
-'5','val_5','5','val_5'
-'5','val_5','5','val_5'
-'8','val_8','8','val_8'
-'9','val_9','9','val_9'
-'tst1','500','tst1','500'
-23 rows selected
->>>
->>> -- test usage of GROUP BY within view
->>> CREATE VIEW view15 AS
-SELECT key,COUNT(value) AS value_count
-FROM src
-GROUP BY key;
-'key','value_count'
-No rows selected
->>> DESCRIBE EXTENDED view15;
-'col_name','data_type','comment'
-'key','string',''
-'value_count','bigint',''
-'','',''
-'Detailed Table Information','Table(tableName:view15, dbName:create_view, owner:!!{user.name}!!, createTime:!!UNIXTIME!!, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value_count, type:bigint, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[], parameters:{transient_lastDdlTime=!!UNIXTIME!!}, viewOriginalText:SELECT key,COUNT(value) AS value_count ',''
-'FROM src ','',''
-'GROUP BY key, viewExpandedText:SELECT `src`.`key`,COUNT(`src`.`value`) AS `value_count` ','',''
-'FROM `create_view`.`src` ','',''
-'GROUP BY `src`.`key`, tableType:VIRTUAL_VIEW)','',''
-8 rows selected
->>> DESCRIBE FORMATTED view15;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'key ','string ','None '
-'value_count ','bigint ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_view ',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Table Type: ','VIRTUAL_VIEW ',''
-'Table Parameters:','',''
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','null ',''
-'InputFormat: ','org.apache.hadoop.mapred.SequenceFileInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'','',''
-'# View Information','',''
-'View Original Text: ','SELECT key,COUNT(value) AS value_count ',''
-'FROM src ','',''
-'GROUP BY key','',''
-'View Expanded Text: ','SELECT `src`.`key`,COUNT(`src`.`value`) AS `value_count` ',''
-'FROM `create_view`.`src` ','',''
-'GROUP BY `src`.`key`','',''
-32 rows selected
->>> SELECT * FROM view15
-ORDER BY value_count DESC, key
-LIMIT 10;
-'key','value_count'
-'230','5'
-'348','5'
-'401','5'
-'469','5'
-'138','4'
-'169','4'
-'277','4'
-'406','4'
-'468','4'
-'489','4'
-10 rows selected
->>>
->>> -- test usage of DISTINCT within view
->>> CREATE VIEW view16 AS
-SELECT DISTINCT value
-FROM src;
-'value'
-No rows selected
->>> DESCRIBE EXTENDED view16;
-'col_name','data_type','comment'
-'value','string',''
-'','',''
-'Detailed Table Information','Table(tableName:view16, dbName:create_view, owner:!!{user.name}!!, createTime:!!UNIXTIME!!, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:value, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[], parameters:{transient_lastDdlTime=!!UNIXTIME!!}, viewOriginalText:SELECT DISTINCT value ',''
-'FROM src, viewExpandedText:SELECT DISTINCT `src`.`value` ','',''
-'FROM `create_view`.`src`, tableType:VIRTUAL_VIEW)','',''
-5 rows selected
->>> DESCRIBE FORMATTED view16;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'value ','string ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_view ',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Table Type: ','VIRTUAL_VIEW ',''
-'Table Parameters:','',''
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','null ',''
-'InputFormat: ','org.apache.hadoop.mapred.SequenceFileInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'','',''
-'# View Information','',''
-'View Original Text: ','SELECT DISTINCT value ',''
-'FROM src','',''
-'View Expanded Text: ','SELECT DISTINCT `src`.`value` ',''
-'FROM `create_view`.`src`','',''
-29 rows selected
->>> SELECT * FROM view16
-ORDER BY value
-LIMIT 10;
-'value'
-'val_0'
-'val_10'
-'val_100'
-'val_103'
-'val_104'
-'val_105'
-'val_11'
-'val_111'
-'val_113'
-'val_114'
-10 rows selected
->>>
->>> -- HIVE-2133: DROP TABLE IF EXISTS should ignore a matching view name
->>> DROP TABLE IF EXISTS view16;
-No rows affected
->>> DESCRIBE view16;
-'col_name','data_type','comment'
-'value','string',''
-1 row selected
->>>
->>> -- Likewise, DROP VIEW IF EXISTS should ignore a matching table name
->>> DROP VIEW IF EXISTS table1;
-No rows affected
->>> DESCRIBE table1;
-'col_name','data_type','comment'
-'key','int',''
-'value','string',''
-2 rows selected
->>>
->>> -- this should work since currently we don't track view->table
->>> -- dependencies for implementing RESTRICT
->>>
->>>
->>> DROP VIEW view1;
-No rows affected
->>> DROP VIEW view2;
-No rows affected
->>> DROP VIEW view3;
-No rows affected
->>> DROP VIEW view4;
-No rows affected
->>> DROP VIEW view5;
-No rows affected
->>> DROP VIEW view6;
-No rows affected
->>> DROP VIEW view7;
-No rows affected
->>> DROP VIEW view8;
-No rows affected
->>> DROP VIEW view9;
-No rows affected
->>> DROP VIEW view10;
-No rows affected
->>> DROP VIEW view11;
-No rows affected
->>> DROP VIEW view12;
-No rows affected
->>> DROP VIEW view13;
-No rows affected
->>> DROP VIEW view14;
-No rows affected
->>> DROP VIEW view15;
-No rows affected
->>> DROP VIEW view16;
-No rows affected
->>> DROP TEMPORARY FUNCTION test_translate;
-No rows affected
->>> DROP TEMPORARY FUNCTION test_max;
-No rows affected
->>> DROP TEMPORARY FUNCTION test_explode;
-No rows affected
->>> !record
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/create_view_partitioned.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/create_view_partitioned.q.out b/ql/src/test/results/beelinepositive/create_view_partitioned.q.out
deleted file mode 100644
index 9460960..0000000
--- a/ql/src/test/results/beelinepositive/create_view_partitioned.q.out
+++ /dev/null
@@ -1,292 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/create_view_partitioned.q.raw". Enter "record" with no arguments to stop it.
->>> !run !!{qFileDirectory}!!/create_view_partitioned.q
->>> DROP VIEW vp1;
-No rows affected
->>> DROP VIEW vp2;
-No rows affected
->>> DROP VIEW vp3;
-No rows affected
->>>
->>> -- test partitioned view definition
->>> -- (underlying table is not actually partitioned)
->>> CREATE VIEW vp1
-PARTITIONED ON (value)
-AS
-SELECT key, value
-FROM src
-WHERE key=86;
-'key','value'
-No rows selected
->>> DESCRIBE EXTENDED vp1;
-'col_name','data_type','comment'
-'key','string',''
-'value','string',''
-'','',''
-'Detailed Table Information','Table(tableName:vp1, dbName:create_view_partitioned, owner:!!{user.name}!!, createTime:!!UNIXTIME!!, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[FieldSchema(name:value, type:string, comment:null)], parameters:{transient_lastDdlTime=!!UNIXTIME!!}, viewOriginalText:SELECT key, value ',''
-'FROM src ','',''
-'WHERE key=86, viewExpandedText:SELECT `src`.`key`, `src`.`value` ','',''
-'FROM `create_view_partitioned`.`src` ','',''
-'WHERE `src`.`key`=86, tableType:VIRTUAL_VIEW)','',''
-8 rows selected
->>> DESCRIBE FORMATTED vp1;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'key ','string ','None '
-'','',''
-'# Partition Information','',''
-'# col_name ','data_type ','comment '
-'','',''
-'value ','string ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_view_partitioned',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Table Type: ','VIRTUAL_VIEW ',''
-'Table Parameters:','',''
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','null ',''
-'InputFormat: ','org.apache.hadoop.mapred.SequenceFileInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'','',''
-'# View Information','',''
-'View Original Text: ','SELECT key, value ',''
-'FROM src ','',''
-'WHERE key=86','',''
-'View Expanded Text: ','SELECT `src`.`key`, `src`.`value` ',''
-'FROM `create_view_partitioned`.`src` ','',''
-'WHERE `src`.`key`=86','',''
-36 rows selected
->>>
->>> SELECT * FROM vp1;
-'key','value'
-'86','val_86'
-1 row selected
->>>
->>> SELECT key FROM vp1;
-'key'
-'86'
-1 row selected
->>>
->>> SELECT value FROM vp1;
-'value'
-'val_86'
-1 row selected
->>>
->>> ALTER VIEW vp1
-ADD PARTITION (value='val_86') PARTITION (value='val_xyz');
-No rows affected
->>>
->>> -- should work since we use IF NOT EXISTS
->>> ALTER VIEW vp1
-ADD IF NOT EXISTS PARTITION (value='val_xyz');
-No rows affected
->>>
->>> SHOW PARTITIONS vp1;
-'partition'
-'value=val_86'
-'value=val_xyz'
-2 rows selected
->>>
->>> SHOW PARTITIONS vp1 PARTITION(value='val_86');
-'partition'
-'value=val_86'
-1 row selected
->>>
->>> SHOW TABLE EXTENDED LIKE vp1;
-'tab_name'
-'tableName:vp1'
-'owner:!!{user.name}!!'
-'location:null'
-'inputformat:org.apache.hadoop.mapred.SequenceFileInputFormat'
-'outputformat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-'columns:struct columns { string key}'
-'partitioned:true'
-'partitionColumns:struct partition_columns { string value}'
-''
-9 rows selected
->>>
->>> SHOW TABLE EXTENDED LIKE vp1 PARTITION(value='val_86');
-'tab_name'
-'tableName:vp1'
-'owner:!!{user.name}!!'
-'location:null'
-'inputformat:org.apache.hadoop.mapred.SequenceFileInputFormat'
-'outputformat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-'columns:struct columns { string key}'
-'partitioned:true'
-'partitionColumns:struct partition_columns { string value}'
-''
-9 rows selected
->>>
->>> ALTER VIEW vp1
-DROP PARTITION (value='val_xyz');
-No rows affected
->>>
->>> SET hive.exec.drop.ignorenonexistent=false;
-No rows affected
->>>
->>> -- should work since we use IF EXISTS
->>> ALTER VIEW vp1
-DROP IF EXISTS PARTITION (value='val_xyz');
-No rows affected
->>>
->>> SHOW PARTITIONS vp1;
-'partition'
-'value=val_86'
-1 row selected
->>>
->>> SET hive.mapred.mode=strict;
-No rows affected
->>>
->>> -- Even though no partition predicate is specified in the next query,
->>> -- the WHERE clause inside of the view should satisfy strict mode.
->>> -- In other words, strict only applies to underlying tables
->>> -- (regardless of whether or not the view is partitioned).
->>> SELECT * FROM vp1;
-'key','value'
-'86','val_86'
-1 row selected
->>>
->>> SET hive.mapred.mode=nonstrict;
-No rows affected
->>>
->>> -- test a partitioned view on top of an underlying partitioned table,
->>> -- but with only a suffix of the partitioning columns
->>> CREATE VIEW vp2
-PARTITIONED ON (hr)
-AS SELECT * FROM srcpart WHERE key < 10;
-'key','value','ds','hr'
-No rows selected
->>> DESCRIBE FORMATTED vp2;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'key ','string ','None '
-'value ','string ','None '
-'ds ','string ','None '
-'','',''
-'# Partition Information','',''
-'# col_name ','data_type ','comment '
-'','',''
-'hr ','string ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_view_partitioned',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Table Type: ','VIRTUAL_VIEW ',''
-'Table Parameters:','',''
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','null ',''
-'InputFormat: ','org.apache.hadoop.mapred.SequenceFileInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'','',''
-'# View Information','',''
-'View Original Text: ','SELECT * FROM srcpart WHERE key < 10',''
-'View Expanded Text: ','SELECT `srcpart`.`key`, `srcpart`.`value`, `srcpart`.`ds`, `srcpart`.`hr` FROM `create_view_partitioned`.`srcpart` WHERE `srcpart`.`key` < 10',''
-34 rows selected
->>>
->>> ALTER VIEW vp2 ADD PARTITION (hr='11') PARTITION (hr='12');
-No rows affected
->>> SELECT key FROM vp2 WHERE hr='12' ORDER BY key;
-'key'
-'0'
-'0'
-'0'
-'0'
-'0'
-'0'
-'2'
-'2'
-'4'
-'4'
-'5'
-'5'
-'5'
-'5'
-'5'
-'5'
-'8'
-'8'
-'9'
-'9'
-20 rows selected
->>>
->>> -- test a partitioned view where the PARTITIONED ON clause references
->>> -- an imposed column name
->>> CREATE VIEW vp3(k,v)
-PARTITIONED ON (v)
-AS
-SELECT key, value
-FROM src
-WHERE key=86;
-'key','value'
-No rows selected
->>> DESCRIBE FORMATTED vp3;
-'col_name','data_type','comment'
-'# col_name ','data_type ','comment '
-'','',''
-'k ','string ','None '
-'','',''
-'# Partition Information','',''
-'# col_name ','data_type ','comment '
-'','',''
-'v ','string ','None '
-'','',''
-'# Detailed Table Information','',''
-'Database: ','create_view_partitioned',''
-'Owner: ','!!{user.name}!! ',''
-'CreateTime: ','!!TIMESTAMP!!',''
-'LastAccessTime: ','UNKNOWN ',''
-'Retention: ','0 ',''
-'Table Type: ','VIRTUAL_VIEW ',''
-'Table Parameters:','',''
-'','transient_lastDdlTime','!!UNIXTIME!! '
-'','',''
-'# Storage Information','',''
-'SerDe Library: ','null ',''
-'InputFormat: ','org.apache.hadoop.mapred.SequenceFileInputFormat',''
-'OutputFormat: ','org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat',''
-'Compressed: ','No ',''
-'Num Buckets: ','-1 ',''
-'Bucket Columns: ','[] ',''
-'Sort Columns: ','[] ',''
-'','',''
-'# View Information','',''
-'View Original Text: ','SELECT key, value ',''
-'FROM src ','',''
-'WHERE key=86','',''
-'View Expanded Text: ','SELECT `key` AS `k`, `value` AS `v` FROM (SELECT `src`.`key`, `src`.`value` ',''
-'FROM `create_view_partitioned`.`src` ','',''
-'WHERE `src`.`key`=86) `vp3`','',''
-36 rows selected
->>>
->>> ALTER VIEW vp3
-ADD PARTITION (v='val_86');
-No rows affected
->>>
->>> DROP VIEW vp1;
-No rows affected
->>> DROP VIEW vp2;
-No rows affected
->>> DROP VIEW vp3;
-No rows affected
->>> !record