You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by na...@apache.org on 2010/01/07 04:02:01 UTC
svn commit: r896744 [3/3] - in /hadoop/hive/trunk: ./
ql/src/java/org/apache/hadoop/hive/ql/exec/
ql/src/java/org/apache/hadoop/hive/ql/udf/
ql/src/java/org/apache/hadoop/hive/ql/udf/generic/
ql/src/test/queries/clientpositive/ ql/src/test/results/clie...
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_unix_timestamp.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_unix_timestamp.q.out?rev=896744&r1=896743&r2=896744&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_unix_timestamp.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_unix_timestamp.q.out Thu Jan 7 03:01:26 2010
@@ -1,17 +1,28 @@
+PREHOOK: query: DESCRIBE FUNCTION unix_timestamp
+PREHOOK: type: DESCFUNCTION
+POSTHOOK: query: DESCRIBE FUNCTION unix_timestamp
+POSTHOOK: type: DESCFUNCTION
+unix_timestamp([date[, pattern]]) - Returns the UNIX timestamp
+PREHOOK: query: DESCRIBE FUNCTION EXTENDED unix_timestamp
+PREHOOK: type: DESCFUNCTION
+POSTHOOK: query: DESCRIBE FUNCTION EXTENDED unix_timestamp
+POSTHOOK: type: DESCFUNCTION
+unix_timestamp([date[, pattern]]) - Returns the UNIX timestamp
+Converts the current or specified time to number of seconds since 1970-01-01.
PREHOOK: query: SELECT
'2009-03-20 11:30:01',
unix_timestamp('2009-03-20 11:30:01')
FROM src LIMIT 1
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/490859379/10000
+PREHOOK: Output: file:/Users/carl/Projects/hd9/hive-trunk/build/ql/tmp/1659161710/10000
POSTHOOK: query: SELECT
'2009-03-20 11:30:01',
unix_timestamp('2009-03-20 11:30:01')
FROM src LIMIT 1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/490859379/10000
+POSTHOOK: Output: file:/Users/carl/Projects/hd9/hive-trunk/build/ql/tmp/1659161710/10000
2009-03-20 11:30:01 1237573801
PREHOOK: query: SELECT
'2009-03-20',
@@ -19,14 +30,14 @@
FROM src LIMIT 1
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1674908517/10000
+PREHOOK: Output: file:/Users/carl/Projects/hd9/hive-trunk/build/ql/tmp/1103088265/10000
POSTHOOK: query: SELECT
'2009-03-20',
unix_timestamp('2009-03-20', 'yyyy-MM-dd')
FROM src LIMIT 1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1674908517/10000
+POSTHOOK: Output: file:/Users/carl/Projects/hd9/hive-trunk/build/ql/tmp/1103088265/10000
2009-03-20 1237532400
PREHOOK: query: SELECT
'2009 Mar 20 11:30:01 am',
@@ -34,14 +45,14 @@
FROM src LIMIT 1
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/265278879/10000
+PREHOOK: Output: file:/Users/carl/Projects/hd9/hive-trunk/build/ql/tmp/1519095937/10000
POSTHOOK: query: SELECT
'2009 Mar 20 11:30:01 am',
unix_timestamp('2009 Mar 20 11:30:01 am', 'yyyy MMM dd h:mm:ss a')
FROM src LIMIT 1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/265278879/10000
+POSTHOOK: Output: file:/Users/carl/Projects/hd9/hive-trunk/build/ql/tmp/1519095937/10000
2009 Mar 20 11:30:01 am 1237573801
PREHOOK: query: SELECT
'random_string',
@@ -49,12 +60,12 @@
FROM src LIMIT 1
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/568226360/10000
+PREHOOK: Output: file:/Users/carl/Projects/hd9/hive-trunk/build/ql/tmp/1021450138/10000
POSTHOOK: query: SELECT
'random_string',
unix_timestamp('random_string')
FROM src LIMIT 1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/568226360/10000
+POSTHOOK: Output: file:/Users/carl/Projects/hd9/hive-trunk/build/ql/tmp/1021450138/10000
random_string NULL
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udtf_explode.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udtf_explode.q.out?rev=896744&r1=896743&r2=896744&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udtf_explode.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udtf_explode.q.out Thu Jan 7 03:01:26 2010
@@ -1,3 +1,13 @@
+PREHOOK: query: DESCRIBE FUNCTION explode
+PREHOOK: type: DESCFUNCTION
+POSTHOOK: query: DESCRIBE FUNCTION explode
+POSTHOOK: type: DESCFUNCTION
+explode(a) - separates the elements of array a into multiple rows
+PREHOOK: query: DESCRIBE FUNCTION EXTENDED explode
+PREHOOK: type: DESCFUNCTION
+POSTHOOK: query: DESCRIBE FUNCTION EXTENDED explode
+POSTHOOK: type: DESCFUNCTION
+explode(a) - separates the elements of array a into multiple rows
PREHOOK: query: EXPLAIN EXTENDED SELECT explode(array(1,2,3)) AS myCol FROM src LIMIT 3
PREHOOK: type: QUERY
POSTHOOK: query: EXPLAIN EXTENDED SELECT explode(array(1,2,3)) AS myCol FROM src LIMIT 3
@@ -27,7 +37,7 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/784241868/10001
+ directory: file:/Users/carl/Projects/hd9/hive-trunk/build/ql/tmp/1666222499/10001
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -37,9 +47,9 @@
columns.types int
Needs Tagging: false
Path -> Alias:
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src [src]
+ file:/Users/carl/Projects/hd9/hive-trunk/build/ql/test/data/warehouse/src [src]
Path -> Partition:
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
+ file:/Users/carl/Projects/hd9/hive-trunk/build/ql/test/data/warehouse/src
Partition
base file name: src
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -54,8 +64,8 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
- transient_lastDdlTime 1262138345
+ location file:/Users/carl/Projects/hd9/hive-trunk/build/ql/test/data/warehouse/src
+ transient_lastDdlTime 1262809225
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -70,8 +80,8 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
- transient_lastDdlTime 1262138345
+ location file:/Users/carl/Projects/hd9/hive-trunk/build/ql/test/data/warehouse/src
+ transient_lastDdlTime 1262809225
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: src
name: src
@@ -116,9 +126,9 @@
type: int
Needs Tagging: false
Path -> Alias:
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src [a:src]
+ file:/Users/carl/Projects/hd9/hive-trunk/build/ql/test/data/warehouse/src [a:src]
Path -> Partition:
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
+ file:/Users/carl/Projects/hd9/hive-trunk/build/ql/test/data/warehouse/src
Partition
base file name: src
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -133,8 +143,8 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
- transient_lastDdlTime 1262138345
+ location file:/Users/carl/Projects/hd9/hive-trunk/build/ql/test/data/warehouse/src
+ transient_lastDdlTime 1262809225
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -149,8 +159,8 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
- transient_lastDdlTime 1262138345
+ location file:/Users/carl/Projects/hd9/hive-trunk/build/ql/test/data/warehouse/src
+ transient_lastDdlTime 1262809225
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: src
name: src
@@ -174,7 +184,7 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/467070148/10002
+ directory: file:/Users/carl/Projects/hd9/hive-trunk/build/ql/tmp/894782310/10002
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -186,7 +196,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/467070148/10002
+ file:/Users/carl/Projects/hd9/hive-trunk/build/ql/tmp/894782310/10002
Reduce Output Operator
key expressions:
expr: _col0
@@ -201,9 +211,9 @@
type: bigint
Needs Tagging: false
Path -> Alias:
- file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/467070148/10002 [file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/467070148/10002]
+ file:/Users/carl/Projects/hd9/hive-trunk/build/ql/tmp/894782310/10002 [file:/Users/carl/Projects/hd9/hive-trunk/build/ql/tmp/894782310/10002]
Path -> Partition:
- file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/467070148/10002
+ file:/Users/carl/Projects/hd9/hive-trunk/build/ql/tmp/894782310/10002
Partition
base file name: 10002
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -239,7 +249,7 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/467070148/10001
+ directory: file:/Users/carl/Projects/hd9/hive-trunk/build/ql/tmp/894782310/10001
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -256,33 +266,33 @@
PREHOOK: query: SELECT explode(array(1,2,3)) AS myCol FROM src LIMIT 3
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1534071913/10000
+PREHOOK: Output: file:/Users/carl/Projects/hd9/hive-trunk/build/ql/tmp/882248436/10000
POSTHOOK: query: SELECT explode(array(1,2,3)) AS myCol FROM src LIMIT 3
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1534071913/10000
+POSTHOOK: Output: file:/Users/carl/Projects/hd9/hive-trunk/build/ql/tmp/882248436/10000
1
2
3
PREHOOK: query: SELECT explode(array(1,2,3)) AS (myCol) FROM src LIMIT 3
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/730345114/10000
+PREHOOK: Output: file:/Users/carl/Projects/hd9/hive-trunk/build/ql/tmp/1127088300/10000
POSTHOOK: query: SELECT explode(array(1,2,3)) AS (myCol) FROM src LIMIT 3
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/730345114/10000
+POSTHOOK: Output: file:/Users/carl/Projects/hd9/hive-trunk/build/ql/tmp/1127088300/10000
1
2
3
PREHOOK: query: SELECT a.myCol, count(1) FROM (SELECT explode(array(1,2,3)) AS myCol FROM src LIMIT 3) a GROUP BY a.myCol
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1456413188/10000
+PREHOOK: Output: file:/Users/carl/Projects/hd9/hive-trunk/build/ql/tmp/1245396324/10000
POSTHOOK: query: SELECT a.myCol, count(1) FROM (SELECT explode(array(1,2,3)) AS myCol FROM src LIMIT 3) a GROUP BY a.myCol
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1456413188/10000
+POSTHOOK: Output: file:/Users/carl/Projects/hd9/hive-trunk/build/ql/tmp/1245396324/10000
1 1
2 1
3 1