You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by zs...@apache.org on 2010/01/24 22:37:42 UTC
svn commit: r902648 - in /hadoop/hive/trunk: ./
ql/src/java/org/apache/hadoop/hive/ql/parse/
ql/src/test/queries/clientnegative/ ql/src/test/queries/clientpositive/
ql/src/test/results/clientnegative/ ql/src/test/results/clientpositive/
Author: zshao
Date: Sun Jan 24 21:37:41 2010
New Revision: 902648
URL: http://svn.apache.org/viewvc?rev=902648&view=rev
Log:
HIVE-1059. Date/DateTime/TimeStamp types should throw an error. (Paul Yang via zshao)
Added:
hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_alter1.q
hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_alter2.q
hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_create1.q
hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_create2.q
hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_create3.q
hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_transform.q
hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_alter1.q.out
hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_alter2.q.out
hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_create1.q.out
hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_create2.q.out
hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_create3.q.out
hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_transform.q.out
Modified:
hadoop/hive/trunk/CHANGES.txt
hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java
hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/ErrorMsg.java
hadoop/hive/trunk/ql/src/test/queries/clientpositive/inputddl2.q
hadoop/hive/trunk/ql/src/test/queries/clientpositive/inputddl4.q
hadoop/hive/trunk/ql/src/test/queries/clientpositive/inputddl6.q
hadoop/hive/trunk/ql/src/test/queries/clientpositive/inputddl8.q
hadoop/hive/trunk/ql/src/test/queries/clientpositive/show_tables.q
hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_create_tbl1.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl2.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl4.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl6.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl8.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/show_tables.q.out
Modified: hadoop/hive/trunk/CHANGES.txt
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/CHANGES.txt?rev=902648&r1=902647&r2=902648&view=diff
==============================================================================
--- hadoop/hive/trunk/CHANGES.txt (original)
+++ hadoop/hive/trunk/CHANGES.txt Sun Jan 24 21:37:41 2010
@@ -501,6 +501,9 @@
HIVE-1072. Keep CreateTime when a partition is overwritten.
(Paul Yang via zshao)
+ HIVE-1059. Date/DateTime/TimeStamp types should throw an error.
+ (Paul Yang via zshao)
+
Release 0.4.0 - Unreleased
INCOMPATIBLE CHANGES
Modified: hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java?rev=902648&r1=902647&r2=902648&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java (original)
+++ hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java Sun Jan 24 21:37:41 2010
@@ -76,7 +76,12 @@
.put(HiveParser.TOK_TIMESTAMP, Constants.TIMESTAMP_TYPE_NAME);
}
- public static String getTypeName(int token) {
+ public static String getTypeName(int token) throws SemanticException {
+ // date, datetime, and timestamp types aren't currently supported
+ if (token == HiveParser.TOK_DATE || token == HiveParser.TOK_DATETIME ||
+ token == HiveParser.TOK_TIMESTAMP ) {
+ throw new SemanticException(ErrorMsg.UNSUPPORTED_TYPE.getMsg());
+ }
return TokenToTypeName.get(token);
}
Modified: hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/ErrorMsg.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/ErrorMsg.java?rev=902648&r1=902647&r2=902648&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/ErrorMsg.java (original)
+++ hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/ErrorMsg.java Sun Jan 24 21:37:41 2010
@@ -118,7 +118,8 @@
"The same output cannot be present multiple times: "), INVALID_AS(
"AS clause has an invalid number of aliases"), VIEW_COL_MISMATCH(
"The number of columns produced by the SELECT clause does not match the number of column names specified by CREATE VIEW"), DML_AGAINST_VIEW(
- "A view cannot be used as target table for LOAD or INSERT");
+ "A view cannot be used as target table for LOAD or INSERT"), UNSUPPORTED_TYPE(
+ "DATE, DATETIME, and TIMESTAMP types aren't supported yet. Please use STRING instead.");
private String mesg;
private String SQLState;
Added: hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_alter1.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_alter1.q?rev=902648&view=auto
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_alter1.q (added)
+++ hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_alter1.q Sun Jan 24 21:37:41 2010
@@ -0,0 +1,2 @@
+CREATE TABLE alter_test (d STRING);
+ALTER TABLE alter_test CHANGE d d DATE;
Added: hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_alter2.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_alter2.q?rev=902648&view=auto
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_alter2.q (added)
+++ hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_alter2.q Sun Jan 24 21:37:41 2010
@@ -0,0 +1,2 @@
+CREATE TABLE alter_test (d STRING);
+ALTER TABLE alter_test ADD COLUMNS (ds DATE);
Added: hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_create1.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_create1.q?rev=902648&view=auto
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_create1.q (added)
+++ hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_create1.q Sun Jan 24 21:37:41 2010
@@ -0,0 +1 @@
+CREATE TABLE date_test (d DATE);
Added: hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_create2.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_create2.q?rev=902648&view=auto
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_create2.q (added)
+++ hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_create2.q Sun Jan 24 21:37:41 2010
@@ -0,0 +1 @@
+CREATE TABLE datetime_test (d DATETIME);
Added: hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_create3.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_create3.q?rev=902648&view=auto
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_create3.q (added)
+++ hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_create3.q Sun Jan 24 21:37:41 2010
@@ -0,0 +1 @@
+CREATE TABLE timestamp_test (d TIMESTAMP);
Added: hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_transform.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_transform.q?rev=902648&view=auto
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_transform.q (added)
+++ hadoop/hive/trunk/ql/src/test/queries/clientnegative/invalid_t_transform.q Sun Jan 24 21:37:41 2010
@@ -0,0 +1 @@
+SELECT TRANSFORM(*) USING 'cat' AS (key DATE) FROM src;
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/inputddl2.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/inputddl2.q?rev=902648&r1=902647&r2=902648&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/inputddl2.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/inputddl2.q Sun Jan 24 21:37:41 2010
@@ -1,6 +1,6 @@
EXPLAIN
-CREATE TABLE INPUTDDL2(key INT, value STRING) PARTITIONED BY(ds DATETIME, country STRING) STORED AS TEXTFILE;
-CREATE TABLE INPUTDDL2(key INT, value STRING) PARTITIONED BY(ds DATETIME, country STRING) STORED AS TEXTFILE;
+CREATE TABLE INPUTDDL2(key INT, value STRING) PARTITIONED BY(ds STRING, country STRING) STORED AS TEXTFILE;
+CREATE TABLE INPUTDDL2(key INT, value STRING) PARTITIONED BY(ds STRING, country STRING) STORED AS TEXTFILE;
DESCRIBE INPUTDDL2;
DROP TABLE INPUTDDL2;
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/inputddl4.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/inputddl4.q?rev=902648&r1=902647&r2=902648&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/inputddl4.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/inputddl4.q Sun Jan 24 21:37:41 2010
@@ -5,7 +5,7 @@
friends ARRAY<BIGINT>, properties MAP<STRING, STRING>,
ip STRING COMMENT 'IP Address of the User')
COMMENT 'This is the page view table'
- PARTITIONED BY(ds DATETIME, country STRING)
+ PARTITIONED BY(ds STRING, country STRING)
CLUSTERED BY(userid) SORTED BY(viewTime) INTO 32 BUCKETS;
DESCRIBE INPUTDDL4;
DESCRIBE EXTENDED INPUTDDL4;
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/inputddl6.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/inputddl6.q?rev=902648&r1=902647&r2=902648&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/inputddl6.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/inputddl6.q Sun Jan 24 21:37:41 2010
@@ -2,7 +2,7 @@
-- test for describe extended table partition
-- test for alter table drop partition
DROP TABLE INPUTDDL6;
-CREATE TABLE INPUTDDL6(KEY STRING, VALUE STRING) PARTITIONED BY(ds DATETIME) STORED AS TEXTFILE;
+CREATE TABLE INPUTDDL6(KEY STRING, VALUE STRING) PARTITIONED BY(ds STRING) STORED AS TEXTFILE;
LOAD DATA LOCAL INPATH '../data/files/kv1.txt' INTO TABLE INPUTDDL6 PARTITION (ds='2008-04-09');
LOAD DATA LOCAL INPATH '../data/files/kv1.txt' INTO TABLE INPUTDDL6 PARTITION (ds='2008-04-08');
DESCRIBE EXTENDED INPUTDDL6;
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/inputddl8.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/inputddl8.q?rev=902648&r1=902647&r2=902648&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/inputddl8.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/inputddl8.q Sun Jan 24 21:37:41 2010
@@ -1,6 +1,6 @@
DROP TABLE INPUTDDL8;
CREATE TABLE INPUTDDL8 COMMENT 'This is a thrift based table'
- PARTITIONED BY(ds DATETIME, country STRING)
+ PARTITIONED BY(ds STRING, country STRING)
CLUSTERED BY(aint) SORTED BY(lint) INTO 32 BUCKETS
ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.thrift.ThriftDeserializer'
WITH SERDEPROPERTIES ('serialization.class' = 'org.apache.hadoop.hive.serde2.thrift.test.Complex',
Modified: hadoop/hive/trunk/ql/src/test/queries/clientpositive/show_tables.q
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/queries/clientpositive/show_tables.q?rev=902648&r1=902647&r2=902648&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/queries/clientpositive/show_tables.q (original)
+++ hadoop/hive/trunk/ql/src/test/queries/clientpositive/show_tables.q Sun Jan 24 21:37:41 2010
@@ -1,5 +1,5 @@
-CREATE TABLE shtb_test1(KEY STRING, VALUE STRING) PARTITIONED BY(ds DATETIME) STORED AS TEXTFILE;
-CREATE TABLE shtb_test2(KEY STRING, VALUE STRING) PARTITIONED BY(ds DATETIME) STORED AS TEXTFILE;
+CREATE TABLE shtb_test1(KEY STRING, VALUE STRING) PARTITIONED BY(ds STRING) STORED AS TEXTFILE;
+CREATE TABLE shtb_test2(KEY STRING, VALUE STRING) PARTITIONED BY(ds STRING) STORED AS TEXTFILE;
EXPLAIN
SHOW TABLES 'shtb_*';
Modified: hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_create_tbl1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_create_tbl1.q.out?rev=902648&r1=902647&r2=902648&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_create_tbl1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_create_tbl1.q.out Sun Jan 24 21:37:41 2010
@@ -2,13 +2,4 @@
PREHOOK: type: DROPTABLE
POSTHOOK: query: DROP TABLE inv_valid_tbl1
POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE inv_valid_tbl1 COMMENT 'This is a thrift based table'
- PARTITIONED BY(aint DATETIME, country STRING)
- CLUSTERED BY(aint) SORTED BY(lint) INTO 32 BUCKETS
- ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.thrift.ThriftDeserializer'
- WITH SERDEPROPERTIES ('serialization.class' = 'org.apache.hadoop.hive.serde2.thrift.test.Complex',
- 'serialization.format' = 'org.apache.thrift.protocol.TBinaryProtocol')
- STORED AS SEQUENCEFILE
-PREHOOK: type: CREATETABLE
-FAILED: Error in metadata: org.apache.hadoop.hive.ql.metadata.HiveException: Partition column name aint conflicts with table columns.
-FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask
+FAILED: Error in semantic analysis: DATE, DATETIME, and TIMESTAMP types aren't supported yet. Please use STRING instead.
Added: hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_alter1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_alter1.q.out?rev=902648&view=auto
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_alter1.q.out (added)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_alter1.q.out Sun Jan 24 21:37:41 2010
@@ -0,0 +1,6 @@
+PREHOOK: query: CREATE TABLE alter_test (d STRING)
+PREHOOK: type: CREATETABLE
+POSTHOOK: query: CREATE TABLE alter_test (d STRING)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: default@alter_test
+FAILED: Error in semantic analysis: DATE, DATETIME, and TIMESTAMP types aren't supported yet. Please use STRING instead.
Added: hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_alter2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_alter2.q.out?rev=902648&view=auto
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_alter2.q.out (added)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_alter2.q.out Sun Jan 24 21:37:41 2010
@@ -0,0 +1,4 @@
+PREHOOK: query: CREATE TABLE alter_test (d STRING)
+PREHOOK: type: CREATETABLE
+FAILED: Error in metadata: AlreadyExistsException(message:Table alter_test already exists)
+FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask
Added: hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_create1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_create1.q.out?rev=902648&view=auto
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_create1.q.out (added)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_create1.q.out Sun Jan 24 21:37:41 2010
@@ -0,0 +1 @@
+FAILED: Error in semantic analysis: DATE, DATETIME, and TIMESTAMP types aren't supported yet. Please use STRING instead.
Added: hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_create2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_create2.q.out?rev=902648&view=auto
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_create2.q.out (added)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_create2.q.out Sun Jan 24 21:37:41 2010
@@ -0,0 +1 @@
+FAILED: Error in semantic analysis: DATE, DATETIME, and TIMESTAMP types aren't supported yet. Please use STRING instead.
Added: hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_create3.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_create3.q.out?rev=902648&view=auto
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_create3.q.out (added)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_create3.q.out Sun Jan 24 21:37:41 2010
@@ -0,0 +1 @@
+FAILED: Error in semantic analysis: DATE, DATETIME, and TIMESTAMP types aren't supported yet. Please use STRING instead.
Added: hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_transform.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_transform.q.out?rev=902648&view=auto
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_transform.q.out (added)
+++ hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_t_transform.q.out Sun Jan 24 21:37:41 2010
@@ -0,0 +1 @@
+FAILED: Error in semantic analysis: DATE, DATETIME, and TIMESTAMP types aren't supported yet. Please use STRING instead.
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl2.q.out?rev=902648&r1=902647&r2=902648&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl2.q.out Sun Jan 24 21:37:41 2010
@@ -1,11 +1,11 @@
PREHOOK: query: EXPLAIN
-CREATE TABLE INPUTDDL2(key INT, value STRING) PARTITIONED BY(ds DATETIME, country STRING) STORED AS TEXTFILE
+CREATE TABLE INPUTDDL2(key INT, value STRING) PARTITIONED BY(ds STRING, country STRING) STORED AS TEXTFILE
PREHOOK: type: CREATETABLE
POSTHOOK: query: EXPLAIN
-CREATE TABLE INPUTDDL2(key INT, value STRING) PARTITIONED BY(ds DATETIME, country STRING) STORED AS TEXTFILE
+CREATE TABLE INPUTDDL2(key INT, value STRING) PARTITIONED BY(ds STRING, country STRING) STORED AS TEXTFILE
POSTHOOK: type: CREATETABLE
ABSTRACT SYNTAX TREE:
- (TOK_CREATETABLE INPUTDDL2 TOK_LIKETABLE (TOK_TABCOLLIST (TOK_TABCOL key TOK_INT) (TOK_TABCOL value TOK_STRING)) (TOK_TABLEPARTCOLS (TOK_TABCOLLIST (TOK_TABCOL ds TOK_DATETIME) (TOK_TABCOL country TOK_STRING))) TOK_TBLTEXTFILE)
+ (TOK_CREATETABLE INPUTDDL2 TOK_LIKETABLE (TOK_TABCOLLIST (TOK_TABCOL key TOK_INT) (TOK_TABCOL value TOK_STRING)) (TOK_TABLEPARTCOLS (TOK_TABCOLLIST (TOK_TABCOL ds TOK_STRING) (TOK_TABCOL country TOK_STRING))) TOK_TBLTEXTFILE)
STAGE DEPENDENCIES:
Stage-0 is a root stage
@@ -19,14 +19,14 @@
input format: org.apache.hadoop.mapred.TextInputFormat
# buckets: -1
output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- partition columns: ds datetime, country string
+ partition columns: ds string, country string
name: INPUTDDL2
isExternal: false
-PREHOOK: query: CREATE TABLE INPUTDDL2(key INT, value STRING) PARTITIONED BY(ds DATETIME, country STRING) STORED AS TEXTFILE
+PREHOOK: query: CREATE TABLE INPUTDDL2(key INT, value STRING) PARTITIONED BY(ds STRING, country STRING) STORED AS TEXTFILE
PREHOOK: type: CREATETABLE
-POSTHOOK: query: CREATE TABLE INPUTDDL2(key INT, value STRING) PARTITIONED BY(ds DATETIME, country STRING) STORED AS TEXTFILE
+POSTHOOK: query: CREATE TABLE INPUTDDL2(key INT, value STRING) PARTITIONED BY(ds STRING, country STRING) STORED AS TEXTFILE
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: default@INPUTDDL2
PREHOOK: query: DESCRIBE INPUTDDL2
@@ -35,7 +35,7 @@
POSTHOOK: type: DESCTABLE
key int
value string
-ds datetime
+ds string
country string
PREHOOK: query: DROP TABLE INPUTDDL2
PREHOOK: type: DROPTABLE
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl4.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl4.q.out?rev=902648&r1=902647&r2=902648&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl4.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl4.q.out Sun Jan 24 21:37:41 2010
@@ -9,7 +9,7 @@
friends ARRAY<BIGINT>, properties MAP<STRING, STRING>,
ip STRING COMMENT 'IP Address of the User')
COMMENT 'This is the page view table'
- PARTITIONED BY(ds DATETIME, country STRING)
+ PARTITIONED BY(ds STRING, country STRING)
CLUSTERED BY(userid) SORTED BY(viewTime) INTO 32 BUCKETS
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE INPUTDDL4(viewTime STRING, userid INT,
@@ -17,7 +17,7 @@
friends ARRAY<BIGINT>, properties MAP<STRING, STRING>,
ip STRING COMMENT 'IP Address of the User')
COMMENT 'This is the page view table'
- PARTITIONED BY(ds DATETIME, country STRING)
+ PARTITIONED BY(ds STRING, country STRING)
CLUSTERED BY(userid) SORTED BY(viewTime) INTO 32 BUCKETS
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: default@INPUTDDL4
@@ -32,7 +32,7 @@
friends array<bigint>
properties map<string,string>
ip string IP Address of the User
-ds datetime
+ds string
country string
PREHOOK: query: DESCRIBE EXTENDED INPUTDDL4
PREHOOK: type: DESCTABLE
@@ -45,10 +45,10 @@
friends array<bigint>
properties map<string,string>
ip string IP Address of the User
-ds datetime
+ds string
country string
-Detailed Table Information Table(tableName:inputddl4, dbName:default, owner:njain, createTime:1253780744, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:viewtime, type:string, comment:null), FieldSchema(name:userid, type:int, comment:null), FieldSchema(name:page_url, type:string, comment:null), FieldSchema(name:referrer_url, type:string, comment:null), FieldSchema(name:friends, type:array<bigint>, comment:null), FieldSchema(name:properties, type:map<string,string>, comment:null), FieldSchema(name:ip, type:string, comment:IP Address of the User)], location:file:/data/users/njain/hive5/hive5/build/ql/test/data/warehouse/inputddl4, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:32, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[userid], sortC
ols:[Order(col:viewtime, order:1)], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:datetime, comment:null), FieldSchema(name:country, type:string, comment:null)], parameters:{comment=This is the page view table})
+Detailed Table Information Table(tableName:inputddl4, dbName:default, owner:pyang, createTime:1264208851, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:viewtime, type:string, comment:null), FieldSchema(name:userid, type:int, comment:null), FieldSchema(name:page_url, type:string, comment:null), FieldSchema(name:referrer_url, type:string, comment:null), FieldSchema(name:friends, type:array<bigint>, comment:null), FieldSchema(name:properties, type:map<string,string>, comment:null), FieldSchema(name:ip, type:string, comment:IP Address of the User)], location:file:/data/users/pyang/task2/trunk/VENDOR.hive/trunk/build/ql/test/data/warehouse/inputddl4, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:32, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCo
ls:[userid], sortCols:[Order(col:viewtime, order:1)], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null), FieldSchema(name:country, type:string, comment:null)], parameters:{transient_lastDdlTime=1264208851,comment=This is the page view table}, viewOriginalText:null, viewExpandedText:null)
PREHOOK: query: DROP TABLE INPUTDDL4
PREHOOK: type: DROPTABLE
POSTHOOK: query: DROP TABLE INPUTDDL4
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl6.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl6.q.out?rev=902648&r1=902647&r2=902648&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl6.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl6.q.out Sun Jan 24 21:37:41 2010
@@ -8,9 +8,9 @@
-- test for alter table drop partition
DROP TABLE INPUTDDL6
POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE INPUTDDL6(KEY STRING, VALUE STRING) PARTITIONED BY(ds DATETIME) STORED AS TEXTFILE
+PREHOOK: query: CREATE TABLE INPUTDDL6(KEY STRING, VALUE STRING) PARTITIONED BY(ds STRING) STORED AS TEXTFILE
PREHOOK: type: CREATETABLE
-POSTHOOK: query: CREATE TABLE INPUTDDL6(KEY STRING, VALUE STRING) PARTITIONED BY(ds DATETIME) STORED AS TEXTFILE
+POSTHOOK: query: CREATE TABLE INPUTDDL6(KEY STRING, VALUE STRING) PARTITIONED BY(ds STRING) STORED AS TEXTFILE
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: default@INPUTDDL6
PREHOOK: query: LOAD DATA LOCAL INPATH '../data/files/kv1.txt' INTO TABLE INPUTDDL6 PARTITION (ds='2008-04-09')
@@ -29,18 +29,18 @@
POSTHOOK: type: DESCTABLE
key string
value string
-ds datetime
+ds string
-Detailed Table Information Table(tableName:inputddl6, dbName:default, owner:njain, createTime:1253780755, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:file:/data/users/njain/hive5/hive5/build/ql/test/data/warehouse/inputddl6, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:datetime, comment:null)], parameters:{})
+Detailed Table Information Table(tableName:inputddl6, dbName:default, owner:pyang, createTime:1264209075, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:file:/data/users/pyang/task2/trunk/VENDOR.hive/trunk/build/ql/test/data/warehouse/inputddl6, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null)], parameters:{transient_lastDdlTime=1264209075}, viewOriginalText:null, viewExpandedText:null)
PREHOOK: query: DESCRIBE EXTENDED INPUTDDL6 PARTITION (ds='2008-04-08')
PREHOOK: type: DESCTABLE
POSTHOOK: query: DESCRIBE EXTENDED INPUTDDL6 PARTITION (ds='2008-04-08')
POSTHOOK: type: DESCTABLE
key string
value string
-ds datetime
+ds string
-Detailed Partition Information Partition(values:[2008-04-08], dbName:default, tableName:inputddl6, createTime:0, lastAccessTime:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:file:/data/users/njain/hive5/hive5/build/ql/test/data/warehouse/inputddl6/ds=2008-04-08, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}), parameters:{})
+Detailed Partition Information Partition(values:[2008-04-08], dbName:default, tableName:inputddl6, createTime:1264209075, lastAccessTime:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:file:/data/users/pyang/task2/trunk/VENDOR.hive/trunk/build/ql/test/data/warehouse/inputddl6/ds=2008-04-08, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}), parameters:{transient_lastDdlTime=1264209075})
PREHOOK: query: SHOW PARTITIONS INPUTDDL6
PREHOOK: type: SHOWPARTITIONS
POSTHOOK: query: SHOW PARTITIONS INPUTDDL6
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl8.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl8.q.out?rev=902648&r1=902647&r2=902648&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl8.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl8.q.out Sun Jan 24 21:37:41 2010
@@ -3,7 +3,7 @@
POSTHOOK: query: DROP TABLE INPUTDDL8
POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE INPUTDDL8 COMMENT 'This is a thrift based table'
- PARTITIONED BY(ds DATETIME, country STRING)
+ PARTITIONED BY(ds STRING, country STRING)
CLUSTERED BY(aint) SORTED BY(lint) INTO 32 BUCKETS
ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.thrift.ThriftDeserializer'
WITH SERDEPROPERTIES ('serialization.class' = 'org.apache.hadoop.hive.serde2.thrift.test.Complex',
@@ -11,7 +11,7 @@
STORED AS SEQUENCEFILE
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE INPUTDDL8 COMMENT 'This is a thrift based table'
- PARTITIONED BY(ds DATETIME, country STRING)
+ PARTITIONED BY(ds STRING, country STRING)
CLUSTERED BY(aint) SORTED BY(lint) INTO 32 BUCKETS
ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.thrift.ThriftDeserializer'
WITH SERDEPROPERTIES ('serialization.class' = 'org.apache.hadoop.hive.serde2.thrift.test.Complex',
@@ -29,10 +29,10 @@
lstring array<string> from deserializer
lintstring array<org.apache.hadoop.hive.serde2.thrift.test.IntString> from deserializer
mstringstring map<string,string> from deserializer
-ds datetime
+ds string
country string
-Detailed Table Information Table(tableName:inputddl8, dbName:default, owner:njain, createTime:1253780774, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[], location:file:/data/users/njain/hive5/hive5/build/ql/test/data/warehouse/inputddl8, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:32, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.thrift.ThriftDeserializer, parameters:{serialization.class=org.apache.hadoop.hive.serde2.thrift.test.Complex,serialization.format=com.facebook.thrift.protocol.TBinaryProtocol}), bucketCols:[aint], sortCols:[Order(col:lint, order:1)], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:datetime, comment:null), FieldSchema(name:country, type:string, comment:null)], parameters:{comment=This is a thrift based table})
+Detailed Table Information Table(tableName:inputddl8, dbName:default, owner:pyang, createTime:1264209638, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[], location:file:/data/users/pyang/task2/trunk/VENDOR.hive/trunk/build/ql/test/data/warehouse/inputddl8, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:32, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.thrift.ThriftDeserializer, parameters:{serialization.class=org.apache.hadoop.hive.serde2.thrift.test.Complex,serialization.format=com.facebook.thrift.protocol.TBinaryProtocol}), bucketCols:[aint], sortCols:[Order(col:lint, order:1)], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null), FieldSchema(name:country, type:string, comment:null)], parameters:{transient_lastDdlTime=1264209638,comment=This is a thrift based table}, viewOriginalText:null,
viewExpandedText:null)
PREHOOK: query: DROP TABLE INPUTDDL8
PREHOOK: type: DROPTABLE
POSTHOOK: query: DROP TABLE INPUTDDL8
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/show_tables.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/show_tables.q.out?rev=902648&r1=902647&r2=902648&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/show_tables.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/show_tables.q.out Sun Jan 24 21:37:41 2010
@@ -1,11 +1,11 @@
-PREHOOK: query: CREATE TABLE shtb_test1(KEY STRING, VALUE STRING) PARTITIONED BY(ds DATETIME) STORED AS TEXTFILE
+PREHOOK: query: CREATE TABLE shtb_test1(KEY STRING, VALUE STRING) PARTITIONED BY(ds STRING) STORED AS TEXTFILE
PREHOOK: type: CREATETABLE
-POSTHOOK: query: CREATE TABLE shtb_test1(KEY STRING, VALUE STRING) PARTITIONED BY(ds DATETIME) STORED AS TEXTFILE
+POSTHOOK: query: CREATE TABLE shtb_test1(KEY STRING, VALUE STRING) PARTITIONED BY(ds STRING) STORED AS TEXTFILE
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: default@shtb_test1
-PREHOOK: query: CREATE TABLE shtb_test2(KEY STRING, VALUE STRING) PARTITIONED BY(ds DATETIME) STORED AS TEXTFILE
+PREHOOK: query: CREATE TABLE shtb_test2(KEY STRING, VALUE STRING) PARTITIONED BY(ds STRING) STORED AS TEXTFILE
PREHOOK: type: CREATETABLE
-POSTHOOK: query: CREATE TABLE shtb_test2(KEY STRING, VALUE STRING) PARTITIONED BY(ds DATETIME) STORED AS TEXTFILE
+POSTHOOK: query: CREATE TABLE shtb_test2(KEY STRING, VALUE STRING) PARTITIONED BY(ds STRING) STORED AS TEXTFILE
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: default@shtb_test2
PREHOOK: query: EXPLAIN