You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by na...@apache.org on 2009/06/15 07:43:29 UTC
svn commit: r784656 [10/30] - in /hadoop/hive/trunk: ./
common/src/java/org/apache/hadoop/hive/conf/
ql/src/java/org/apache/hadoop/hive/ql/exec/
ql/src/java/org/apache/hadoop/hive/ql/optimizer/
ql/src/java/org/apache/hadoop/hive/ql/parse/ ql/src/java/o...
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out Mon Jun 15 05:43:22 2009
@@ -32,14 +32,14 @@
type: string
Select Operator
expressions:
- expr: UDFToInteger(0)
+ expr: UDFToInteger(_col0)
type: int
- expr: 1
+ expr: _col1
type: string
File Output Operator
compressed: false
GlobalTableId: 1
- directory: /data/users/zshao/tools/495-trunk-apache-hive/ql/../build/ql/tmp/1044305705/895861142.10000.insclause-0
+ directory: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/358216208/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -53,14 +53,14 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/zshao/tools/495-trunk-apache-hive/build/ql/test/data/warehouse/dest1
+ location file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/dest1
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest1
Needs Tagging: false
Path -> Alias:
- file:/data/users/zshao/tools/495-trunk-apache-hive/build/ql/test/data/warehouse/srcbucket/kv1.txt
+ file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcbucket/kv1.txt
Path -> Partition:
- file:/data/users/zshao/tools/495-trunk-apache-hive/build/ql/test/data/warehouse/srcbucket/kv1.txt
+ file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcbucket/kv1.txt
Partition
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -76,7 +76,7 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/zshao/tools/495-trunk-apache-hive/build/ql/test/data/warehouse/srcbucket
+ location file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcbucket
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcbucket
@@ -84,7 +84,7 @@
Move Operator
tables:
replace: true
- source: /data/users/zshao/tools/495-trunk-apache-hive/ql/../build/ql/tmp/1044305705/895861142.10000.insclause-0
+ source: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/358216208/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -98,10 +98,10 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/zshao/tools/495-trunk-apache-hive/build/ql/test/data/warehouse/dest1
+ location file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/dest1
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest1
- tmp directory: /data/users/zshao/tools/495-trunk-apache-hive/ql/../build/ql/tmp/1044305705/895861142.10001
+ tmp directory: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/358216208/10001
query: INSERT OVERWRITE TABLE dest1 SELECT s.*
@@ -111,7 +111,7 @@
Output: default/dest1
query: SELECT dest1.* FROM dest1
Input: default/dest1
-Output: /data/users/zshao/tools/495-trunk-apache-hive/ql/../build/ql/tmp/578063712/363033982.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1723450216/10000
165 val_165
484 val_484
150 val_150
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/sample8.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/sample8.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/sample8.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/sample8.q.out Mon Jun 15 05:43:22 2009
@@ -19,24 +19,22 @@
Map Reduce
Alias -> Map Operator Tree:
t
- Select Operator
- expressions:
- expr: key
- type: string
- expr: value
- type: string
- Filter Operator
- predicate:
- expr: (((hash(0) & 2147483647) % 10) = 0)
- type: boolean
- Reduce Output Operator
- sort order:
- tag: 1
- value expressions:
- expr: 0
- type: string
- expr: 1
- type: string
+ Filter Operator
+ predicate:
+ expr: (((hash(key) & 2147483647) % 10) = 0)
+ type: boolean
+ Reduce Output Operator
+ sort order:
+ tag: 1
+ value expressions:
+ expr: key
+ type: string
+ expr: value
+ type: string
+ expr: ds
+ type: string
+ expr: hr
+ type: string
s
Filter Operator
predicate:
@@ -56,12 +54,12 @@
type: string
Needs Tagging: true
Path -> Alias:
- file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
- file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
- file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11
- file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12
+ file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
+ file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
+ file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11
+ file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12
Path -> Partition:
- file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
+ file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
Partition
partition values:
ds 2008-04-08
@@ -80,10 +78,10 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart
+ location file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
- file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
+ file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
Partition
partition values:
ds 2008-04-08
@@ -102,10 +100,10 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart
+ location file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
- file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11
+ file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11
Partition
partition values:
ds 2008-04-09
@@ -124,10 +122,10 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart
+ location file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
- file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12
+ file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12
Partition
partition values:
ds 2008-04-09
@@ -146,7 +144,7 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart
+ location file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
Reduce Operator Tree:
@@ -154,73 +152,73 @@
condition map:
Inner Join 0 to 1
condition expressions:
- 0 {VALUE.0} {VALUE.1} {VALUE.2} {VALUE.3}
- 1 {VALUE.0} {VALUE.1}
+ 0 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3}
+ 1 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3}
Filter Operator
predicate:
- expr: ((((((4 = 0) and (5 = 1)) and (2 = '2008-04-08')) and (3 = '11')) and (2 = '2008-04-08')) and (3 = '11'))
+ expr: ((((((_col4 = _col0) and (_col5 = _col1)) and (_col2 = '2008-04-08')) and (_col3 = '11')) and (_col2 = '2008-04-08')) and (_col3 = '11'))
type: boolean
Select Operator
expressions:
- expr: 0
+ expr: _col0
type: string
- expr: 1
+ expr: _col1
type: string
- expr: 2
+ expr: _col2
type: string
- expr: 3
+ expr: _col3
type: string
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/njain/hive2/hive2/build/ql/tmp/1443892627/10004
+ directory: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1858927587/10002
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
properties:
name binary_table
- serialization.ddl struct binary_table { string temporarycol0, string temporarycol1, string temporarycol2, string temporarycol3}
+ serialization.ddl struct binary_table { string _col0, string _col1, string _col2, string _col3}
serialization.format com.facebook.thrift.protocol.TBinaryProtocol
name: binary_table
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/njain/hive2/hive2/build/ql/tmp/1443892627/10004
+ file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1858927587/10002
Reduce Output Operator
key expressions:
- expr: 0
+ expr: _col0
type: string
- expr: 1
+ expr: _col1
type: string
sort order: ++
Map-reduce partition columns:
- expr: 0
+ expr: _col0
type: string
- expr: 1
+ expr: _col1
type: string
tag: -1
value expressions:
- expr: 0
+ expr: _col0
type: string
- expr: 1
+ expr: _col1
type: string
- expr: 2
+ expr: _col2
type: string
- expr: 3
+ expr: _col3
type: string
Needs Tagging: false
Path -> Alias:
- file:/data/users/njain/hive2/hive2/build/ql/tmp/1443892627/10004
+ file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1858927587/10002
Path -> Partition:
- file:/data/users/njain/hive2/hive2/build/ql/tmp/1443892627/10004
+ file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1858927587/10002
Partition
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
properties:
name binary_table
- serialization.ddl struct binary_table { string temporarycol0, string temporarycol1, string temporarycol2, string temporarycol3}
+ serialization.ddl struct binary_table { string _col0, string _col1, string _col2, string _col3}
serialization.format com.facebook.thrift.protocol.TBinaryProtocol
name: binary_table
Reduce Operator Tree:
@@ -228,7 +226,7 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/njain/hive2/hive2/build/ql/tmp/1443892627/10003
+ directory: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1858927587/10001
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -251,7 +249,7 @@
Input: default/srcpart/ds=2008-04-08/hr=12
Input: default/srcpart/ds=2008-04-09/hr=11
Input: default/srcpart/ds=2008-04-09/hr=12
-Output: file:/data/users/njain/hive2/hive2/build/ql/tmp/1318003535/10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/587483226/10000
0 val_0
0 val_0
0 val_0
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/sort.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/sort.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/sort.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/sort.q.out Mon Jun 15 05:43:22 2009
@@ -20,14 +20,14 @@
type: string
Reduce Output Operator
key expressions:
- expr: 0
+ expr: _col0
type: string
sort order: +
tag: -1
value expressions:
- expr: 0
+ expr: _col0
type: string
- expr: 1
+ expr: _col1
type: string
Reduce Operator Tree:
Extract
@@ -45,7 +45,7 @@
query: SELECT x.* FROM SRC x SORT BY key
Input: default/src
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/875507913/8639603.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1617011148/10000
0 val_0
0 val_0
0 val_0
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/subq.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/subq.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/subq.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/subq.q.out Mon Jun 15 05:43:22 2009
@@ -27,9 +27,9 @@
type: string
Select Operator
expressions:
- expr: 0
+ expr: _col0
type: string
- expr: 1
+ expr: _col1
type: string
File Output Operator
compressed: false
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out Mon Jun 15 05:43:22 2009
@@ -22,44 +22,44 @@
aggregations:
expr: count(1)
keys:
- expr: 0
+ expr: key
type: string
mode: hash
Reduce Output Operator
key expressions:
- expr: 0
+ expr: _col0
type: string
sort order: +
Map-reduce partition columns:
- expr: 0
+ expr: _col0
type: string
tag: -1
value expressions:
- expr: 1
+ expr: _col1
type: bigint
Reduce Operator Tree:
Group By Operator
aggregations:
- expr: count(VALUE.0)
+ expr: count(VALUE._col0)
keys:
- expr: KEY.0
+ expr: KEY._col0
type: string
mode: mergepartial
Select Operator
expressions:
- expr: 0
+ expr: _col0
type: string
- expr: 1
+ expr: _col1
type: bigint
Filter Operator
predicate:
- expr: (UDFToDouble(0) >= UDFToDouble(90))
+ expr: (UDFToDouble(_col0) >= UDFToDouble(90))
type: boolean
Select Operator
expressions:
- expr: 0
+ expr: _col0
type: string
- expr: 1
+ expr: _col1
type: bigint
File Output Operator
compressed: false
@@ -77,7 +77,7 @@
FROM (SELECT b.key as k, count(1) as c FROM src b GROUP BY b.key) a
WHERE a.k >= 90
Input: default/src
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/900838886/139366063.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/2044684588/10000
100 2
103 2
104 2
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf1.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf1.q.out Mon Jun 15 05:43:22 2009
@@ -23,64 +23,60 @@
Map Reduce
Alias -> Map Operator Tree:
src
- Select Operator
- expressions:
- expr: key
- type: string
- Filter Operator
- predicate:
- expr: (UDFToDouble(0) = UDFToDouble(86))
- type: boolean
- Select Operator
- expressions:
- expr: ('a' like '%a%')
- type: boolean
- expr: ('b' like '%a%')
- type: boolean
- expr: ('ab' like '%a%')
- type: boolean
- expr: ('ab' like '%a_')
- type: boolean
- expr: ('%_' like '\%\_')
- type: boolean
- expr: ('ab' like '\%\_')
- type: boolean
- expr: ('ab' like '_a%')
- type: boolean
- expr: ('ab' like 'a')
- type: boolean
- expr: ('' regexp '.*')
- type: boolean
- expr: ('a' regexp '[ab]')
- type: boolean
- expr: ('' regexp '[ab]')
- type: boolean
- expr: ('hadoop' regexp '[a-z]*')
- type: boolean
- expr: ('hadoop' regexp 'o*')
- type: boolean
- expr: regexp_replace('abc', 'b', 'c')
- type: string
- expr: regexp_replace('abc', 'z', 'a')
- type: string
- expr: regexp_replace('abbbb', 'bb', 'b')
- type: string
- expr: regexp_replace('hadoop', '(.)[a-z]*', '$1ive')
- type: string
- expr: regexp_replace('hadoopAAA', 'A.*', '')
- type: string
- expr: regexp_replace('abc', '', 'A')
- type: string
- expr: ('abc' regexp '')
- type: boolean
- File Output Operator
- compressed: false
- GlobalTableId: 1
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: dest1
+ Filter Operator
+ predicate:
+ expr: (UDFToDouble(key) = UDFToDouble(86))
+ type: boolean
+ Select Operator
+ expressions:
+ expr: ('a' like '%a%')
+ type: boolean
+ expr: ('b' like '%a%')
+ type: boolean
+ expr: ('ab' like '%a%')
+ type: boolean
+ expr: ('ab' like '%a_')
+ type: boolean
+ expr: ('%_' like '\%\_')
+ type: boolean
+ expr: ('ab' like '\%\_')
+ type: boolean
+ expr: ('ab' like '_a%')
+ type: boolean
+ expr: ('ab' like 'a')
+ type: boolean
+ expr: ('' regexp '.*')
+ type: boolean
+ expr: ('a' regexp '[ab]')
+ type: boolean
+ expr: ('' regexp '[ab]')
+ type: boolean
+ expr: ('hadoop' regexp '[a-z]*')
+ type: boolean
+ expr: ('hadoop' regexp 'o*')
+ type: boolean
+ expr: regexp_replace('abc', 'b', 'c')
+ type: string
+ expr: regexp_replace('abc', 'z', 'a')
+ type: string
+ expr: regexp_replace('abbbb', 'bb', 'b')
+ type: string
+ expr: regexp_replace('hadoop', '(.)[a-z]*', '$1ive')
+ type: string
+ expr: regexp_replace('hadoopAAA', 'A.*', '')
+ type: string
+ expr: regexp_replace('abc', '', 'A')
+ type: string
+ expr: ('abc' regexp '')
+ type: boolean
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: dest1
Stage: Stage-0
Move Operator
@@ -104,5 +100,5 @@
Output: default/dest1
query: SELECT dest1.* FROM dest1
Input: default/dest1
-Output: /data/users/zshao/tools/495-trunk-apache-hive/ql/../build/ql/tmp/572903932/16944307.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/516885779/10000
true false true true true false false false true true false true false acc abc abb hive hadoop AaAbAcA false
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf3.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf3.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf3.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf3.q.out Mon Jun 15 05:43:22 2009
@@ -27,36 +27,36 @@
sort order:
tag: -1
value expressions:
- expr: 0
+ expr: _col0
type: bigint
- expr: 1
+ expr: _col1
type: double
- expr: 2
+ expr: _col2
type: string
- expr: 3
+ expr: _col3
type: int
- expr: 4
+ expr: _col4
type: int
Reduce Operator Tree:
Group By Operator
aggregations:
- expr: count(VALUE.0)
- expr: sum(VALUE.1)
- expr: avg(VALUE.2)
- expr: min(VALUE.3)
- expr: max(VALUE.4)
+ expr: count(VALUE._col0)
+ expr: sum(VALUE._col1)
+ expr: avg(VALUE._col2)
+ expr: min(VALUE._col3)
+ expr: max(VALUE._col4)
mode: mergepartial
Select Operator
expressions:
- expr: 0
+ expr: _col0
type: bigint
- expr: 1
+ expr: _col1
type: double
- expr: 2
+ expr: _col2
type: double
- expr: 3
+ expr: _col3
type: int
- expr: 4
+ expr: _col4
type: int
File Output Operator
compressed: false
@@ -84,5 +84,5 @@
Output: default/dest1
query: SELECT dest1.* FROM dest1
Input: default/dest1
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/441448894/123433014.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1041569072/10000
0 NULL NULL NULL NULL
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf4.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf4.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf4.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf4.q.out Mon Jun 15 05:43:22 2009
@@ -17,52 +17,51 @@
Alias -> Map Operator Tree:
dest1
Select Operator
- Select Operator
- expressions:
- expr: round(1.0)
- type: bigint
- expr: round(1.5)
- type: bigint
- expr: round(- 1.5)
- type: bigint
- expr: floor(1.0)
- type: bigint
- expr: floor(1.5)
- type: bigint
- expr: floor(- 1.5)
- type: bigint
- expr: sqrt(1.0)
- type: double
- expr: sqrt(- 1.0)
- type: double
- expr: sqrt(0.0)
- type: double
- expr: ceiling(1.0)
- type: bigint
- expr: ceiling(1.5)
- type: bigint
- expr: ceiling(- 1.5)
- type: bigint
- expr: ceiling(1.0)
- type: bigint
- expr: rand(UDFToLong(3))
- type: double
- expr: 3
- type: int
- expr: - 3
- type: int
- expr: (1 + 2)
- type: int
- expr: (1 + - 2)
- type: int
- expr: ~ 1
- type: int
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ expressions:
+ expr: round(1.0)
+ type: bigint
+ expr: round(1.5)
+ type: bigint
+ expr: round(- 1.5)
+ type: bigint
+ expr: floor(1.0)
+ type: bigint
+ expr: floor(1.5)
+ type: bigint
+ expr: floor(- 1.5)
+ type: bigint
+ expr: sqrt(1.0)
+ type: double
+ expr: sqrt(- 1.0)
+ type: double
+ expr: sqrt(0.0)
+ type: double
+ expr: ceiling(1.0)
+ type: bigint
+ expr: ceiling(1.5)
+ type: bigint
+ expr: ceiling(- 1.5)
+ type: bigint
+ expr: ceiling(1.0)
+ type: bigint
+ expr: rand(UDFToLong(3))
+ type: double
+ expr: 3
+ type: int
+ expr: - 3
+ type: int
+ expr: (1 + 2)
+ type: int
+ expr: (1 + - 2)
+ type: int
+ expr: ~ 1
+ type: int
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Stage: Stage-0
Fetch Operator
@@ -71,5 +70,5 @@
query: SELECT round(1.0), round(1.5), round(-1.5), floor(1.0), floor(1.5), floor(-1.5), sqrt(1.0), sqrt(-1.0), sqrt(0.0), ceil(1.0), ceil(1.5), ceil(-1.5), ceiling(1.0), rand(3), +3, -3, 1++2, 1+-2, ~1 FROM dest1
Input: default/dest1
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/29787657/1016730635.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1468603599/10000
1 2 -2 1 1 -2 1.0 NULL 0.0 1 2 -1 1 0.731057369148862 3 -3 3 -1 -2
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf5.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf5.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf5.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf5.q.out Mon Jun 15 05:43:22 2009
@@ -17,30 +17,29 @@
Alias -> Map Operator Tree:
dest1
Select Operator
- Select Operator
- expressions:
- expr: from_unixtime(1226446340)
- type: string
- expr: to_date(from_unixtime(1226446340))
- type: string
- expr: day('2008-11-01')
- type: int
- expr: month('2008-11-01')
- type: int
- expr: year('2008-11-01')
- type: int
- expr: day('2008-11-01 15:32:20')
- type: int
- expr: month('2008-11-01 15:32:20')
- type: int
- expr: year('2008-11-01 15:32:20')
- type: int
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ expressions:
+ expr: from_unixtime(1226446340)
+ type: string
+ expr: to_date(from_unixtime(1226446340))
+ type: string
+ expr: day('2008-11-01')
+ type: int
+ expr: month('2008-11-01')
+ type: int
+ expr: year('2008-11-01')
+ type: int
+ expr: day('2008-11-01 15:32:20')
+ type: int
+ expr: month('2008-11-01 15:32:20')
+ type: int
+ expr: year('2008-11-01 15:32:20')
+ type: int
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Stage: Stage-0
Fetch Operator
@@ -49,5 +48,5 @@
query: SELECT from_unixtime(1226446340), to_date(from_unixtime(1226446340)), day('2008-11-01'), month('2008-11-01'), year('2008-11-01'), day('2008-11-01 15:32:20'), month('2008-11-01 15:32:20'), year('2008-11-01 15:32:20') FROM dest1
Input: default/dest1
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/396140468/103691975.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/467850962/10000
2008-11-11 15:32:20 2008-11-11 1 11 2008 1 11 2008
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf6.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf6.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf6.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf6.q.out Mon Jun 15 05:43:22 2009
@@ -17,16 +17,15 @@
Alias -> Map Operator Tree:
dest1
Select Operator
- Select Operator
- expressions:
- expr: if true, 1, 2
- type: int
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ expressions:
+ expr: if true, 1, 2
+ type: int
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Stage: Stage-0
Fetch Operator
@@ -35,7 +34,7 @@
query: SELECT IF(TRUE, 1, 2) FROM dest1
Input: default/dest1
-Output: /data/users/zshao/tools/495-trunk-apache-hive/ql/../build/ql/tmp/185617439/895542101.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/923424840/10000
1
query: EXPLAIN
SELECT IF(TRUE, 1, 2), IF(FALSE, 1, 2), IF(NULL, 1, 2), IF(TRUE, "a", "b"),
@@ -56,36 +55,35 @@
Alias -> Map Operator Tree:
dest1
Select Operator
- Select Operator
- expressions:
- expr: if true, 1, 2
- type: int
- expr: if false, 1, 2
- type: int
- expr: if null, 1, 2
- type: int
- expr: if true, 'a', 'b'
- type: string
- expr: if true, 0.1, 0.2
- type: double
- expr: if false, UDFToLong(1), UDFToLong(2)
- type: bigint
- expr: if false, UDFToInteger(UDFToByte(127)), 128
- type: int
- expr: if false, UDFToShort(UDFToByte(127)), UDFToShort(128)
- type: smallint
- expr: 128
- type: int
- expr: 1.0
- type: double
- expr: '128'
- type: string
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ expressions:
+ expr: if true, 1, 2
+ type: int
+ expr: if false, 1, 2
+ type: int
+ expr: if null, 1, 2
+ type: int
+ expr: if true, 'a', 'b'
+ type: string
+ expr: if true, 0.1, 0.2
+ type: double
+ expr: if false, UDFToLong(1), UDFToLong(2)
+ type: bigint
+ expr: if false, UDFToInteger(UDFToByte(127)), 128
+ type: int
+ expr: if false, UDFToShort(UDFToByte(127)), UDFToShort(128)
+ type: smallint
+ expr: 128
+ type: int
+ expr: 1.0
+ type: double
+ expr: '128'
+ type: string
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Stage: Stage-0
Fetch Operator
@@ -98,5 +96,5 @@
CAST(128 AS SMALLINT)), CAST(128 AS INT), CAST(1.0 AS DOUBLE),
CAST('128' AS STRING) FROM dest1
Input: default/dest1
-Output: /data/users/zshao/tools/495-trunk-apache-hive/ql/../build/ql/tmp/238198611/144848990.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1340342481/10000
1 2 2 a 0.1 2 128 128 128 1.0 128
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf7.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf7.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf7.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf7.q.out Mon Jun 15 05:43:22 2009
@@ -22,64 +22,63 @@
Alias -> Map Operator Tree:
dest1
Select Operator
- Select Operator
- expressions:
- expr: round(ln(3.0), 12)
- type: double
- expr: ln(0.0)
- type: double
- expr: ln(UDFToDouble(- 1))
- type: double
- expr: round(log(3.0), 12)
- type: double
- expr: log(0.0)
- type: double
- expr: log(UDFToDouble(- 1))
- type: double
- expr: round(log2(3.0), 12)
- type: double
- expr: log2(0.0)
- type: double
- expr: log2(UDFToDouble(- 1))
- type: double
- expr: round(log10(3.0), 12)
- type: double
- expr: log10(0.0)
- type: double
- expr: log10(UDFToDouble(- 1))
- type: double
- expr: round(log(UDFToDouble(2), 3.0), 12)
- type: double
- expr: log(UDFToDouble(2), 0.0)
- type: double
- expr: log(UDFToDouble(2), UDFToDouble(- 1))
- type: double
- expr: log(0.5, UDFToDouble(2))
- type: double
- expr: log(UDFToDouble(2), 0.5)
- type: double
- expr: round(exp(2.0), 12)
- type: double
- expr: pow(UDFToDouble(2), UDFToDouble(3))
- type: double
- expr: pow(UDFToDouble(2), UDFToDouble(3))
- type: double
- expr: pow(UDFToDouble(2), UDFToDouble(- 3))
- type: double
- expr: pow(0.5, UDFToDouble(- 3))
- type: double
- expr: pow(UDFToDouble(4), 0.5)
- type: double
- expr: pow(UDFToDouble(- 1), 0.5)
- type: double
- expr: pow(UDFToDouble(- 1), UDFToDouble(2))
- type: double
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ expressions:
+ expr: round(ln(3.0), 12)
+ type: double
+ expr: ln(0.0)
+ type: double
+ expr: ln(UDFToDouble(- 1))
+ type: double
+ expr: round(log(3.0), 12)
+ type: double
+ expr: log(0.0)
+ type: double
+ expr: log(UDFToDouble(- 1))
+ type: double
+ expr: round(log2(3.0), 12)
+ type: double
+ expr: log2(0.0)
+ type: double
+ expr: log2(UDFToDouble(- 1))
+ type: double
+ expr: round(log10(3.0), 12)
+ type: double
+ expr: log10(0.0)
+ type: double
+ expr: log10(UDFToDouble(- 1))
+ type: double
+ expr: round(log(UDFToDouble(2), 3.0), 12)
+ type: double
+ expr: log(UDFToDouble(2), 0.0)
+ type: double
+ expr: log(UDFToDouble(2), UDFToDouble(- 1))
+ type: double
+ expr: log(0.5, UDFToDouble(2))
+ type: double
+ expr: log(UDFToDouble(2), 0.5)
+ type: double
+ expr: round(exp(2.0), 12)
+ type: double
+ expr: pow(UDFToDouble(2), UDFToDouble(3))
+ type: double
+ expr: pow(UDFToDouble(2), UDFToDouble(3))
+ type: double
+ expr: pow(UDFToDouble(2), UDFToDouble(- 3))
+ type: double
+ expr: pow(0.5, UDFToDouble(- 3))
+ type: double
+ expr: pow(UDFToDouble(4), 0.5)
+ type: double
+ expr: pow(UDFToDouble(- 1), 0.5)
+ type: double
+ expr: pow(UDFToDouble(- 1), UDFToDouble(2))
+ type: double
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Stage: Stage-0
Fetch Operator
@@ -93,5 +92,5 @@
POW(2,3), POWER(2,3), POWER(2,-3), POWER(0.5, -3), POWER(4, 0.5),
POWER(-1, 0.5), POWER(-1, 2) FROM dest1
Input: default/dest1
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/323872951/193616753.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/987082819/10000
1.098612288668 NULL NULL 1.098612288668 NULL NULL 1.584962500721 NULL NULL 0.47712125472 NULL NULL 1.584962500721 NULL NULL NULL -1.0 7.389056098931 8.0 8.0 0.125 8.0 2.0 NaN 1.0
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf8.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf8.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf8.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf8.q.out Mon Jun 15 05:43:22 2009
@@ -19,36 +19,40 @@
Map Reduce
Alias -> Map Operator Tree:
dest1
- Group By Operator
- aggregations:
- expr: avg(UDFToDouble(c1))
- expr: sum(UDFToDouble(c1))
- expr: count(c1)
- mode: hash
- Reduce Output Operator
- sort order:
- tag: -1
- value expressions:
- expr: 0
- type: string
- expr: 1
- type: double
- expr: 2
- type: bigint
+ Select Operator
+ expressions:
+ expr: c1
+ type: string
+ Group By Operator
+ aggregations:
+ expr: avg(UDFToDouble(c1))
+ expr: sum(UDFToDouble(c1))
+ expr: count(c1)
+ mode: hash
+ Reduce Output Operator
+ sort order:
+ tag: -1
+ value expressions:
+ expr: _col0
+ type: string
+ expr: _col1
+ type: double
+ expr: _col2
+ type: bigint
Reduce Operator Tree:
Group By Operator
aggregations:
- expr: avg(VALUE.0)
- expr: sum(VALUE.1)
- expr: count(VALUE.2)
+ expr: avg(VALUE._col0)
+ expr: sum(VALUE._col1)
+ expr: count(VALUE._col2)
mode: mergepartial
Select Operator
expressions:
- expr: 0
+ expr: _col0
type: double
- expr: 1
+ expr: _col1
type: double
- expr: 2
+ expr: _col2
type: bigint
File Output Operator
compressed: false
@@ -64,5 +68,5 @@
query: SELECT avg(c1), sum(c1), count(c1) FROM dest1
Input: default/dest1
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/148346145/763466676.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1344361217/10000
1.0 1.0 1
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf9.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf9.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf9.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf9.q.out Mon Jun 15 05:43:22 2009
@@ -20,54 +20,50 @@
Map Reduce
Alias -> Map Operator Tree:
src
- Select Operator
- expressions:
- expr: key
- type: string
- Filter Operator
- predicate:
- expr: (UDFToDouble(0) = UDFToDouble(86))
- type: boolean
- Select Operator
- expressions:
- expr: datediff('2008-12-31', '2009-01-01')
- type: int
- expr: datediff('2008-03-01', '2008-02-28')
- type: int
- expr: datediff('2007-03-01', '2007-01-28')
- type: int
- expr: datediff('2008-03-01 23:59:59', '2008-03-02 00:00:00')
- type: int
- expr: date_add('2008-12-31', 1)
- type: string
- expr: date_add('2008-12-31', 365)
- type: string
- expr: date_add('2008-02-28', 2)
- type: string
- expr: date_add('2009-02-28', 2)
- type: string
- expr: date_add('2007-02-28', 365)
- type: string
- expr: date_add('2007-02-28 23:59:59', 730)
- type: string
- expr: date_sub('2009-01-01', 1)
- type: string
- expr: date_sub('2009-01-01', 365)
- type: string
- expr: date_sub('2008-02-28', 2)
- type: string
- expr: date_sub('2009-02-28', 2)
- type: string
- expr: date_sub('2007-02-28', 365)
- type: string
- expr: date_sub('2007-02-28 01:12:34', 730)
- type: string
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ Filter Operator
+ predicate:
+ expr: (UDFToDouble(key) = UDFToDouble(86))
+ type: boolean
+ Select Operator
+ expressions:
+ expr: datediff('2008-12-31', '2009-01-01')
+ type: int
+ expr: datediff('2008-03-01', '2008-02-28')
+ type: int
+ expr: datediff('2007-03-01', '2007-01-28')
+ type: int
+ expr: datediff('2008-03-01 23:59:59', '2008-03-02 00:00:00')
+ type: int
+ expr: date_add('2008-12-31', 1)
+ type: string
+ expr: date_add('2008-12-31', 365)
+ type: string
+ expr: date_add('2008-02-28', 2)
+ type: string
+ expr: date_add('2009-02-28', 2)
+ type: string
+ expr: date_add('2007-02-28', 365)
+ type: string
+ expr: date_add('2007-02-28 23:59:59', 730)
+ type: string
+ expr: date_sub('2009-01-01', 1)
+ type: string
+ expr: date_sub('2009-01-01', 365)
+ type: string
+ expr: date_sub('2008-02-28', 2)
+ type: string
+ expr: date_sub('2009-02-28', 2)
+ type: string
+ expr: date_sub('2007-02-28', 365)
+ type: string
+ expr: date_sub('2007-02-28 01:12:34', 730)
+ type: string
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Stage: Stage-0
Fetch Operator
@@ -84,5 +80,5 @@
DATE_SUB('2007-02-28', 365), DATE_SUB('2007-02-28 01:12:34', 730)
FROM src WHERE src.key = 86
Input: default/src
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/804319999/376709074.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/2099135137/10000
-1 2 32 -1 2009-01-01 2009-12-31 2008-03-01 2009-03-02 2008-02-28 2009-02-27 2008-12-31 2008-01-02 2008-02-28 2009-02-27 2006-02-28 2005-02-28
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_10_trims.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_10_trims.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_10_trims.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_10_trims.q.out Mon Jun 15 05:43:22 2009
@@ -16,26 +16,22 @@
Map Reduce
Alias -> Map Operator Tree:
src
- Select Operator
- expressions:
- expr: key
- type: string
- Filter Operator
- predicate:
- expr: (UDFToDouble(0) = UDFToDouble(86))
- type: boolean
- Select Operator
- expressions:
- expr: trim(trim(trim(trim(trim(trim(trim(trim(trim(trim(' abc '))))))))))
- type: string
- File Output Operator
- compressed: false
- GlobalTableId: 1
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: dest1
+ Filter Operator
+ predicate:
+ expr: (UDFToDouble(key) = UDFToDouble(86))
+ type: boolean
+ Select Operator
+ expressions:
+ expr: trim(trim(trim(trim(trim(trim(trim(trim(trim(trim(' abc '))))))))))
+ type: string
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: dest1
Stage: Stage-0
Move Operator
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case.q.out Mon Jun 15 05:43:22 2009
@@ -38,27 +38,26 @@
Alias -> Map Operator Tree:
src
Select Operator
- Select Operator
- expressions:
- expr: CASE (1) WHEN (1) THEN (2) WHEN (3) THEN (4) ELSE (5) END
- type: int
- expr: CASE (2) WHEN (1) THEN (2) ELSE (5) END
- type: int
- expr: CASE (14) WHEN (12) THEN (13) WHEN (14) THEN (15) END
- type: int
- expr: CASE (16) WHEN (12) THEN (13) WHEN (14) THEN (15) END
- type: int
- expr: CASE (17) WHEN (18) THEN (null) WHEN (17) THEN (20) END
- type: int
- expr: CASE (21) WHEN (22) THEN (23) WHEN (21) THEN (24) END
- type: int
- Limit
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ expressions:
+ expr: CASE (1) WHEN (1) THEN (2) WHEN (3) THEN (4) ELSE (5) END
+ type: int
+ expr: CASE (2) WHEN (1) THEN (2) ELSE (5) END
+ type: int
+ expr: CASE (14) WHEN (12) THEN (13) WHEN (14) THEN (15) END
+ type: int
+ expr: CASE (16) WHEN (12) THEN (13) WHEN (14) THEN (15) END
+ type: int
+ expr: CASE (17) WHEN (18) THEN (null) WHEN (17) THEN (20) END
+ type: int
+ expr: CASE (21) WHEN (22) THEN (23) WHEN (21) THEN (24) END
+ type: int
+ Limit
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Stage: Stage-0
Fetch Operator
@@ -92,5 +91,5 @@
END
FROM src LIMIT 1
Input: default/src
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/116746963/155290781.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1131510224/10000
2 5 15 NULL 20 24
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case_column_pruning.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case_column_pruning.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case_column_pruning.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case_column_pruning.q.out Mon Jun 15 05:43:22 2009
@@ -20,49 +20,45 @@
Map Reduce
Alias -> Map Operator Tree:
b
- Select Operator
- expressions:
+ Reduce Output Operator
+ key expressions:
+ expr: key
+ type: string
+ sort order: +
+ Map-reduce partition columns:
expr: key
type: string
- Reduce Output Operator
- key expressions:
- expr: 0
- type: string
- sort order: +
- Map-reduce partition columns:
- expr: 0
- type: string
- tag: 1
- value expressions:
- expr: 0
- type: string
+ tag: 1
+ value expressions:
+ expr: key
+ type: string
+ expr: value
+ type: string
a
- Select Operator
- expressions:
+ Reduce Output Operator
+ key expressions:
+ expr: key
+ type: string
+ sort order: +
+ Map-reduce partition columns:
expr: key
type: string
- Reduce Output Operator
- key expressions:
- expr: 0
- type: string
- sort order: +
- Map-reduce partition columns:
- expr: 0
- type: string
- tag: 0
- value expressions:
- expr: 0
- type: string
+ tag: 0
+ value expressions:
+ expr: key
+ type: string
+ expr: value
+ type: string
Reduce Operator Tree:
Join Operator
condition map:
Inner Join 0 to 1
condition expressions:
- 0 {VALUE.0}
- 1 {VALUE.0}
+ 0 {VALUE._col0} {VALUE._col1}
+ 1 {VALUE._col0} {VALUE._col1}
Select Operator
expressions:
- expr: CASE (0) WHEN ('1') THEN (2) WHEN ('3') THEN (4) ELSE (5) END
+ expr: CASE (_col0) WHEN ('1') THEN (2) WHEN ('3') THEN (4) ELSE (5) END
type: int
File Output Operator
compressed: false
@@ -75,15 +71,15 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/njain/hive2/hive2/build/ql/tmp/982397491/10004
+ file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1081703636/10002
Reduce Output Operator
key expressions:
- expr: 0
+ expr: _col0
type: int
sort order: +
tag: -1
value expressions:
- expr: 0
+ expr: _col0
type: int
Reduce Operator Tree:
Extract
@@ -109,7 +105,7 @@
ON a.key = b.key
ORDER BY key LIMIT 10
Input: default/src
-Output: file:/data/users/njain/hive2/hive2/build/ql/tmp/163558385/10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/10034618/10000
5
5
5
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case_thrift.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case_thrift.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case_thrift.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case_thrift.q.out Mon Jun 15 05:43:22 2009
@@ -29,25 +29,19 @@
src_thrift
Select Operator
expressions:
- expr: lint
- type: array<int>
- expr: lstring
- type: array<string>
- Select Operator
- expressions:
- expr: CASE (0[0]) WHEN (0) THEN ((0[0] + 1)) WHEN (1) THEN ((0[0] + 2)) WHEN (2) THEN (100) ELSE (5) END
- type: int
- expr: CASE (1[0]) WHEN ('0') THEN ('zero') WHEN ('10') THEN (concat(1[0], ' is ten')) ELSE ('default') END
- type: string
- expr: CASE (1[0]) WHEN ('0') THEN (1) ELSE (null) END[0]
- type: string
- Limit
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ expr: CASE (lint[0]) WHEN (0) THEN ((lint[0] + 1)) WHEN (1) THEN ((lint[0] + 2)) WHEN (2) THEN (100) ELSE (5) END
+ type: int
+ expr: CASE (lstring[0]) WHEN ('0') THEN ('zero') WHEN ('10') THEN (concat(lstring[0], ' is ten')) ELSE ('default') END
+ type: string
+ expr: CASE (lstring[0]) WHEN ('0') THEN (lstring) ELSE (null) END[0]
+ type: string
+ Limit
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Stage: Stage-0
Fetch Operator
@@ -71,7 +65,7 @@
END)[0]
FROM src_thrift LIMIT 3
Input: default/src_thrift
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/1401002108/1327666619.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1872084071/10000
1 zero 0
3 10 is ten NULL
100 default NULL
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_coalesce.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_coalesce.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_coalesce.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_coalesce.q.out Mon Jun 15 05:43:22 2009
@@ -31,51 +31,50 @@
Alias -> Map Operator Tree:
src
Select Operator
- Select Operator
- expressions:
- expr: COALESCE(1)
- type: int
- expr: COALESCE(1,2)
- type: int
- expr: COALESCE(null,2)
- type: int
- expr: COALESCE(1,null)
- type: int
- expr: COALESCE(null,null,3)
- type: int
- expr: COALESCE(4,null,null,null)
- type: int
- expr: COALESCE('1')
- type: string
- expr: COALESCE('1','2')
- type: string
- expr: COALESCE(null,'2')
- type: string
- expr: COALESCE('1',null)
- type: string
- expr: COALESCE(null,null,'3')
- type: string
- expr: COALESCE('4',null,null,null)
- type: string
- expr: COALESCE(1.0)
- type: double
- expr: COALESCE(1.0,2.0)
- type: double
- expr: COALESCE(null,2.0)
- type: double
- expr: COALESCE(null,2.0,3.0)
- type: double
- expr: COALESCE(2.0,null,3.0)
- type: double
- expr: COALESCE(if true, null, 0,null)
- type: int
- Limit
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ expressions:
+ expr: COALESCE(1)
+ type: int
+ expr: COALESCE(1,2)
+ type: int
+ expr: COALESCE(null,2)
+ type: int
+ expr: COALESCE(1,null)
+ type: int
+ expr: COALESCE(null,null,3)
+ type: int
+ expr: COALESCE(4,null,null,null)
+ type: int
+ expr: COALESCE('1')
+ type: string
+ expr: COALESCE('1','2')
+ type: string
+ expr: COALESCE(null,'2')
+ type: string
+ expr: COALESCE('1',null)
+ type: string
+ expr: COALESCE(null,null,'3')
+ type: string
+ expr: COALESCE('4',null,null,null)
+ type: string
+ expr: COALESCE(1.0)
+ type: double
+ expr: COALESCE(1.0,2.0)
+ type: double
+ expr: COALESCE(null,2.0)
+ type: double
+ expr: COALESCE(null,2.0,3.0)
+ type: double
+ expr: COALESCE(2.0,null,3.0)
+ type: double
+ expr: COALESCE(if true, null, 0,null)
+ type: int
+ Limit
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Stage: Stage-0
Fetch Operator
@@ -102,7 +101,7 @@
COALESCE(IF(TRUE, NULL, 0), NULL)
FROM src LIMIT 1
Input: default/src
-Output: file:/data/users/zshao/tools/495-trunk-apache-hive/build/ql/tmp/64965527/10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1424226650/10000
1 1 2 1 3 4 1 1 2 1 3 4 1.0 1.0 2.0 2.0 2.0 NULL
query: EXPLAIN
SELECT COALESCE(src_thrift.lint[1], 999),
@@ -123,26 +122,18 @@
src_thrift
Select Operator
expressions:
- expr: lint
- type: array<int>
- expr: lintstring
- type: array<struct<myint:int,mystring:string,underscore_int:int>>
- expr: mstringstring
- type: map<string,string>
- Select Operator
- expressions:
- expr: COALESCE(0[1],999)
- type: int
- expr: COALESCE(1[0].mystring,'999')
- type: string
- expr: COALESCE(2['key_2'],'999')
- type: string
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ expr: COALESCE(lint[1],999)
+ type: int
+ expr: COALESCE(lintstring[0].mystring,'999')
+ type: string
+ expr: COALESCE(mstringstring['key_2'],'999')
+ type: string
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Stage: Stage-0
Fetch Operator
@@ -154,7 +145,7 @@
COALESCE(src_thrift.mstringstring['key_2'], '999')
FROM src_thrift
Input: default/src_thrift
-Output: file:/data/users/zshao/tools/495-trunk-apache-hive/build/ql/tmp/2056249139/10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/300664928/10000
0 0 999
2 1 999
4 8 value_2
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_hash.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_hash.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_hash.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_hash.q.out Mon Jun 15 05:43:22 2009
@@ -18,37 +18,36 @@
Alias -> Map Operator Tree:
src
Select Operator
- Select Operator
- expressions:
- expr: hash(UDFToByte(1))
- type: int
- expr: hash(UDFToShort(2))
- type: int
- expr: hash(3)
- type: int
- expr: hash(UDFToLong('123456789012'))
- type: int
- expr: hash(UDFToFloat(1.25))
- type: int
- expr: hash(16.0)
- type: int
- expr: hash('400')
- type: int
- expr: hash('abc')
- type: int
- expr: hash(true)
- type: int
- expr: hash(false)
- type: int
- expr: hash(1,2,3)
- type: int
- Limit
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ expressions:
+ expr: hash(UDFToByte(1))
+ type: int
+ expr: hash(UDFToShort(2))
+ type: int
+ expr: hash(3)
+ type: int
+ expr: hash(UDFToLong('123456789012'))
+ type: int
+ expr: hash(UDFToFloat(1.25))
+ type: int
+ expr: hash(16.0)
+ type: int
+ expr: hash('400')
+ type: int
+ expr: hash('abc')
+ type: int
+ expr: hash(true)
+ type: int
+ expr: hash(false)
+ type: int
+ expr: hash(1,2,3)
+ type: int
+ Limit
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Stage: Stage-0
Fetch Operator
@@ -62,5 +61,5 @@
hash(1, 2, 3)
FROM src LIMIT 1
Input: default/src
-Output: /data/users/zshao/tools/495-trunk-apache-hive/ql/../build/ql/tmp/347375001/1065831137.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1354544735/10000
1 2 3 -1097262584 1067450368 1076887552 51508 96354 1 0 1026
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_length.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_length.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_length.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_length.q.out Mon Jun 15 05:43:22 2009
@@ -15,20 +15,16 @@
src
Select Operator
expressions:
- expr: value
- type: string
- Select Operator
- expressions:
- expr: length(0)
- type: int
- File Output Operator
- compressed: false
- GlobalTableId: 1
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: dest1
+ expr: length(value)
+ type: int
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: dest1
Stage: Stage-0
Move Operator
@@ -46,7 +42,7 @@
Output: default/dest1
query: SELECT dest1.* FROM dest1
Input: default/dest1
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/432923810/244566699.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/423777377/10000
7
6
7
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_lower.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_lower.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_lower.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_lower.q.out Mon Jun 15 05:43:22 2009
@@ -12,26 +12,22 @@
Map Reduce
Alias -> Map Operator Tree:
src
- Select Operator
- expressions:
- expr: key
- type: string
- Filter Operator
- predicate:
- expr: (UDFToDouble(0) = UDFToDouble(86))
- type: boolean
- Select Operator
- expressions:
- expr: lower('AbC 123')
- type: string
- expr: ucase('AbC 123')
- type: string
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ Filter Operator
+ predicate:
+ expr: (UDFToDouble(key) = UDFToDouble(86))
+ type: boolean
+ Select Operator
+ expressions:
+ expr: lower('AbC 123')
+ type: string
+ expr: ucase('AbC 123')
+ type: string
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Stage: Stage-0
Fetch Operator
@@ -40,5 +36,5 @@
query: SELECT lower('AbC 123'), upper('AbC 123') FROM src WHERE key = 86
Input: default/src
-Output: /data/users/zshao/tools/495-trunk-apache-hive/ql/../build/ql/tmp/1060690797/116583439.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1926782866/10000
abc 123 ABC 123
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_when.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_when.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_when.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_when.q.out Mon Jun 15 05:43:22 2009
@@ -38,27 +38,26 @@
Alias -> Map Operator Tree:
src
Select Operator
- Select Operator
- expressions:
- expr: CASE WHEN ((1 = 1)) THEN (2) WHEN ((1 = 3)) THEN (4) ELSE (5) END
- type: int
- expr: CASE WHEN ((6 = 7)) THEN (8) ELSE (9) END
- type: int
- expr: CASE WHEN ((10 = 11)) THEN (12) WHEN ((13 = 13)) THEN (14) END
- type: int
- expr: CASE WHEN ((15 = 16)) THEN (17) WHEN ((18 = 19)) THEN (20) END
- type: int
- expr: CASE WHEN ((21 = 22)) THEN (null) WHEN ((23 = 23)) THEN (24) END
- type: int
- expr: CASE WHEN ((25 = 26)) THEN (27) WHEN ((28 = 28)) THEN (null) END
- type: int
- Limit
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ expressions:
+ expr: CASE WHEN ((1 = 1)) THEN (2) WHEN ((1 = 3)) THEN (4) ELSE (5) END
+ type: int
+ expr: CASE WHEN ((6 = 7)) THEN (8) ELSE (9) END
+ type: int
+ expr: CASE WHEN ((10 = 11)) THEN (12) WHEN ((13 = 13)) THEN (14) END
+ type: int
+ expr: CASE WHEN ((15 = 16)) THEN (17) WHEN ((18 = 19)) THEN (20) END
+ type: int
+ expr: CASE WHEN ((21 = 22)) THEN (null) WHEN ((23 = 23)) THEN (24) END
+ type: int
+ expr: CASE WHEN ((25 = 26)) THEN (27) WHEN ((28 = 28)) THEN (null) END
+ type: int
+ Limit
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Stage: Stage-0
Fetch Operator
@@ -92,5 +91,5 @@
END
FROM src LIMIT 1
Input: default/src
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/262454053/648350622.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/855982416/10000
2 9 14 NULL 24 NULL
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/union.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/union.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/union.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/union.q.out Mon Jun 15 05:43:22 2009
@@ -32,9 +32,9 @@
Union
Select Operator
expressions:
- expr: 0
+ expr: _col0
type: string
- expr: 1
+ expr: _col1
type: string
File Output Operator
compressed: false
@@ -56,9 +56,9 @@
Union
Select Operator
expressions:
- expr: 0
+ expr: _col0
type: string
- expr: 1
+ expr: _col1
type: string
File Output Operator
compressed: false