You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by na...@apache.org on 2009/06/15 07:43:29 UTC

svn commit: r784656 [10/30] - in /hadoop/hive/trunk: ./ common/src/java/org/apache/hadoop/hive/conf/ ql/src/java/org/apache/hadoop/hive/ql/exec/ ql/src/java/org/apache/hadoop/hive/ql/optimizer/ ql/src/java/org/apache/hadoop/hive/ql/parse/ ql/src/java/o...

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out Mon Jun 15 05:43:22 2009
@@ -32,14 +32,14 @@
                         type: string
                   Select Operator
                     expressions:
-                          expr: UDFToInteger(0)
+                          expr: UDFToInteger(_col0)
                           type: int
-                          expr: 1
+                          expr: _col1
                           type: string
                     File Output Operator
                       compressed: false
                       GlobalTableId: 1
-                      directory: /data/users/zshao/tools/495-trunk-apache-hive/ql/../build/ql/tmp/1044305705/895861142.10000.insclause-0
+                      directory: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/358216208/10000
                       table:
                           input format: org.apache.hadoop.mapred.TextInputFormat
                           output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -53,14 +53,14 @@
                             serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                             file.inputformat org.apache.hadoop.mapred.TextInputFormat
                             file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                            location file:/data/users/zshao/tools/495-trunk-apache-hive/build/ql/test/data/warehouse/dest1
+                            location file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/dest1
                           serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                           name: dest1
       Needs Tagging: false
       Path -> Alias:
-        file:/data/users/zshao/tools/495-trunk-apache-hive/build/ql/test/data/warehouse/srcbucket/kv1.txt 
+        file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcbucket/kv1.txt 
       Path -> Partition:
-        file:/data/users/zshao/tools/495-trunk-apache-hive/build/ql/test/data/warehouse/srcbucket/kv1.txt 
+        file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcbucket/kv1.txt 
           Partition
           
               input format: org.apache.hadoop.mapred.TextInputFormat
@@ -76,7 +76,7 @@
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/zshao/tools/495-trunk-apache-hive/build/ql/test/data/warehouse/srcbucket
+                location file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcbucket
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: srcbucket
 
@@ -84,7 +84,7 @@
     Move Operator
       tables:
             replace: true
-            source: /data/users/zshao/tools/495-trunk-apache-hive/ql/../build/ql/tmp/1044305705/895861142.10000.insclause-0
+            source: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/358216208/10000
             table:
                 input format: org.apache.hadoop.mapred.TextInputFormat
                 output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -98,10 +98,10 @@
                   serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                   file.inputformat org.apache.hadoop.mapred.TextInputFormat
                   file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  location file:/data/users/zshao/tools/495-trunk-apache-hive/build/ql/test/data/warehouse/dest1
+                  location file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/dest1
                 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                 name: dest1
-            tmp directory: /data/users/zshao/tools/495-trunk-apache-hive/ql/../build/ql/tmp/1044305705/895861142.10001
+            tmp directory: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/358216208/10001
 
 
 query: INSERT OVERWRITE TABLE dest1 SELECT s.* 
@@ -111,7 +111,7 @@
 Output: default/dest1
 query: SELECT dest1.* FROM dest1
 Input: default/dest1
-Output: /data/users/zshao/tools/495-trunk-apache-hive/ql/../build/ql/tmp/578063712/363033982.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1723450216/10000
 165	val_165
 484	val_484
 150	val_150

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/sample8.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/sample8.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/sample8.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/sample8.q.out Mon Jun 15 05:43:22 2009
@@ -19,24 +19,22 @@
     Map Reduce
       Alias -> Map Operator Tree:
         t 
-            Select Operator
-              expressions:
-                    expr: key
-                    type: string
-                    expr: value
-                    type: string
-              Filter Operator
-                predicate:
-                    expr: (((hash(0) & 2147483647) % 10) = 0)
-                    type: boolean
-                Reduce Output Operator
-                  sort order: 
-                  tag: 1
-                  value expressions:
-                        expr: 0
-                        type: string
-                        expr: 1
-                        type: string
+            Filter Operator
+              predicate:
+                  expr: (((hash(key) & 2147483647) % 10) = 0)
+                  type: boolean
+              Reduce Output Operator
+                sort order: 
+                tag: 1
+                value expressions:
+                      expr: key
+                      type: string
+                      expr: value
+                      type: string
+                      expr: ds
+                      type: string
+                      expr: hr
+                      type: string
         s 
             Filter Operator
               predicate:
@@ -56,12 +54,12 @@
                       type: string
       Needs Tagging: true
       Path -> Alias:
-        file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 
-        file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 
-        file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11 
-        file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12 
+        file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 
+        file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 
+        file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11 
+        file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12 
       Path -> Partition:
-        file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 
+        file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 
           Partition
             partition values:
               ds 2008-04-08
@@ -80,10 +78,10 @@
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart
+                location file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: srcpart
-        file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 
+        file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 
           Partition
             partition values:
               ds 2008-04-08
@@ -102,10 +100,10 @@
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart
+                location file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: srcpart
-        file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11 
+        file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11 
           Partition
             partition values:
               ds 2008-04-09
@@ -124,10 +122,10 @@
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart
+                location file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: srcpart
-        file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12 
+        file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12 
           Partition
             partition values:
               ds 2008-04-09
@@ -146,7 +144,7 @@
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/njain/hive2/hive2/build/ql/test/data/warehouse/srcpart
+                location file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: srcpart
       Reduce Operator Tree:
@@ -154,73 +152,73 @@
           condition map:
                Inner Join 0 to 1
           condition expressions:
-            0 {VALUE.0} {VALUE.1} {VALUE.2} {VALUE.3}
-            1 {VALUE.0} {VALUE.1}
+            0 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3}
+            1 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3}
           Filter Operator
             predicate:
-                expr: ((((((4 = 0) and (5 = 1)) and (2 = '2008-04-08')) and (3 = '11')) and (2 = '2008-04-08')) and (3 = '11'))
+                expr: ((((((_col4 = _col0) and (_col5 = _col1)) and (_col2 = '2008-04-08')) and (_col3 = '11')) and (_col2 = '2008-04-08')) and (_col3 = '11'))
                 type: boolean
             Select Operator
               expressions:
-                    expr: 0
+                    expr: _col0
                     type: string
-                    expr: 1
+                    expr: _col1
                     type: string
-                    expr: 2
+                    expr: _col2
                     type: string
-                    expr: 3
+                    expr: _col3
                     type: string
               File Output Operator
                 compressed: false
                 GlobalTableId: 0
-                directory: file:/data/users/njain/hive2/hive2/build/ql/tmp/1443892627/10004
+                directory: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1858927587/10002
                 table:
                     input format: org.apache.hadoop.mapred.SequenceFileInputFormat
                     output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
                     properties:
                       name binary_table
-                      serialization.ddl struct binary_table { string temporarycol0, string temporarycol1, string temporarycol2, string temporarycol3}
+                      serialization.ddl struct binary_table { string _col0, string _col1, string _col2, string _col3}
                       serialization.format com.facebook.thrift.protocol.TBinaryProtocol
                     name: binary_table
 
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/data/users/njain/hive2/hive2/build/ql/tmp/1443892627/10004 
+        file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1858927587/10002 
             Reduce Output Operator
               key expressions:
-                    expr: 0
+                    expr: _col0
                     type: string
-                    expr: 1
+                    expr: _col1
                     type: string
               sort order: ++
               Map-reduce partition columns:
-                    expr: 0
+                    expr: _col0
                     type: string
-                    expr: 1
+                    expr: _col1
                     type: string
               tag: -1
               value expressions:
-                    expr: 0
+                    expr: _col0
                     type: string
-                    expr: 1
+                    expr: _col1
                     type: string
-                    expr: 2
+                    expr: _col2
                     type: string
-                    expr: 3
+                    expr: _col3
                     type: string
       Needs Tagging: false
       Path -> Alias:
-        file:/data/users/njain/hive2/hive2/build/ql/tmp/1443892627/10004 
+        file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1858927587/10002 
       Path -> Partition:
-        file:/data/users/njain/hive2/hive2/build/ql/tmp/1443892627/10004 
+        file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1858927587/10002 
           Partition
           
               input format: org.apache.hadoop.mapred.SequenceFileInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
               properties:
                 name binary_table
-                serialization.ddl struct binary_table { string temporarycol0, string temporarycol1, string temporarycol2, string temporarycol3}
+                serialization.ddl struct binary_table { string _col0, string _col1, string _col2, string _col3}
                 serialization.format com.facebook.thrift.protocol.TBinaryProtocol
               name: binary_table
       Reduce Operator Tree:
@@ -228,7 +226,7 @@
           File Output Operator
             compressed: false
             GlobalTableId: 0
-            directory: file:/data/users/njain/hive2/hive2/build/ql/tmp/1443892627/10003
+            directory: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1858927587/10001
             table:
                 input format: org.apache.hadoop.mapred.TextInputFormat
                 output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -251,7 +249,7 @@
 Input: default/srcpart/ds=2008-04-08/hr=12
 Input: default/srcpart/ds=2008-04-09/hr=11
 Input: default/srcpart/ds=2008-04-09/hr=12
-Output: file:/data/users/njain/hive2/hive2/build/ql/tmp/1318003535/10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/587483226/10000
 0	val_0
 0	val_0
 0	val_0

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/sort.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/sort.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/sort.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/sort.q.out Mon Jun 15 05:43:22 2009
@@ -20,14 +20,14 @@
                     type: string
               Reduce Output Operator
                 key expressions:
-                      expr: 0
+                      expr: _col0
                       type: string
                 sort order: +
                 tag: -1
                 value expressions:
-                      expr: 0
+                      expr: _col0
                       type: string
-                      expr: 1
+                      expr: _col1
                       type: string
       Reduce Operator Tree:
         Extract
@@ -45,7 +45,7 @@
 
 query: SELECT x.* FROM SRC x SORT BY key
 Input: default/src
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/875507913/8639603.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1617011148/10000
 0	val_0
 0	val_0
 0	val_0

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/subq.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/subq.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/subq.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/subq.q.out Mon Jun 15 05:43:22 2009
@@ -27,9 +27,9 @@
                       type: string
                 Select Operator
                   expressions:
-                        expr: 0
+                        expr: _col0
                         type: string
-                        expr: 1
+                        expr: _col1
                         type: string
                   File Output Operator
                     compressed: false

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out Mon Jun 15 05:43:22 2009
@@ -22,44 +22,44 @@
                 aggregations:
                       expr: count(1)
                 keys:
-                      expr: 0
+                      expr: key
                       type: string
                 mode: hash
                 Reduce Output Operator
                   key expressions:
-                        expr: 0
+                        expr: _col0
                         type: string
                   sort order: +
                   Map-reduce partition columns:
-                        expr: 0
+                        expr: _col0
                         type: string
                   tag: -1
                   value expressions:
-                        expr: 1
+                        expr: _col1
                         type: bigint
       Reduce Operator Tree:
         Group By Operator
           aggregations:
-                expr: count(VALUE.0)
+                expr: count(VALUE._col0)
           keys:
-                expr: KEY.0
+                expr: KEY._col0
                 type: string
           mode: mergepartial
           Select Operator
             expressions:
-                  expr: 0
+                  expr: _col0
                   type: string
-                  expr: 1
+                  expr: _col1
                   type: bigint
             Filter Operator
               predicate:
-                  expr: (UDFToDouble(0) >= UDFToDouble(90))
+                  expr: (UDFToDouble(_col0) >= UDFToDouble(90))
                   type: boolean
               Select Operator
                 expressions:
-                      expr: 0
+                      expr: _col0
                       type: string
-                      expr: 1
+                      expr: _col1
                       type: bigint
                 File Output Operator
                   compressed: false
@@ -77,7 +77,7 @@
 FROM (SELECT b.key as k, count(1) as c FROM src b GROUP BY b.key) a
 WHERE a.k >= 90
 Input: default/src
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/900838886/139366063.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/2044684588/10000
 100	2
 103	2
 104	2

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf1.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf1.q.out Mon Jun 15 05:43:22 2009
@@ -23,64 +23,60 @@
     Map Reduce
       Alias -> Map Operator Tree:
         src 
-            Select Operator
-              expressions:
-                    expr: key
-                    type: string
-              Filter Operator
-                predicate:
-                    expr: (UDFToDouble(0) = UDFToDouble(86))
-                    type: boolean
-                Select Operator
-                  expressions:
-                        expr: ('a' like '%a%')
-                        type: boolean
-                        expr: ('b' like '%a%')
-                        type: boolean
-                        expr: ('ab' like '%a%')
-                        type: boolean
-                        expr: ('ab' like '%a_')
-                        type: boolean
-                        expr: ('%_' like '\%\_')
-                        type: boolean
-                        expr: ('ab' like '\%\_')
-                        type: boolean
-                        expr: ('ab' like '_a%')
-                        type: boolean
-                        expr: ('ab' like 'a')
-                        type: boolean
-                        expr: ('' regexp '.*')
-                        type: boolean
-                        expr: ('a' regexp '[ab]')
-                        type: boolean
-                        expr: ('' regexp '[ab]')
-                        type: boolean
-                        expr: ('hadoop' regexp '[a-z]*')
-                        type: boolean
-                        expr: ('hadoop' regexp 'o*')
-                        type: boolean
-                        expr: regexp_replace('abc', 'b', 'c')
-                        type: string
-                        expr: regexp_replace('abc', 'z', 'a')
-                        type: string
-                        expr: regexp_replace('abbbb', 'bb', 'b')
-                        type: string
-                        expr: regexp_replace('hadoop', '(.)[a-z]*', '$1ive')
-                        type: string
-                        expr: regexp_replace('hadoopAAA', 'A.*', '')
-                        type: string
-                        expr: regexp_replace('abc', '', 'A')
-                        type: string
-                        expr: ('abc' regexp '')
-                        type: boolean
-                  File Output Operator
-                    compressed: false
-                    GlobalTableId: 1
-                    table:
-                        input format: org.apache.hadoop.mapred.TextInputFormat
-                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                        name: dest1
+            Filter Operator
+              predicate:
+                  expr: (UDFToDouble(key) = UDFToDouble(86))
+                  type: boolean
+              Select Operator
+                expressions:
+                      expr: ('a' like '%a%')
+                      type: boolean
+                      expr: ('b' like '%a%')
+                      type: boolean
+                      expr: ('ab' like '%a%')
+                      type: boolean
+                      expr: ('ab' like '%a_')
+                      type: boolean
+                      expr: ('%_' like '\%\_')
+                      type: boolean
+                      expr: ('ab' like '\%\_')
+                      type: boolean
+                      expr: ('ab' like '_a%')
+                      type: boolean
+                      expr: ('ab' like 'a')
+                      type: boolean
+                      expr: ('' regexp '.*')
+                      type: boolean
+                      expr: ('a' regexp '[ab]')
+                      type: boolean
+                      expr: ('' regexp '[ab]')
+                      type: boolean
+                      expr: ('hadoop' regexp '[a-z]*')
+                      type: boolean
+                      expr: ('hadoop' regexp 'o*')
+                      type: boolean
+                      expr: regexp_replace('abc', 'b', 'c')
+                      type: string
+                      expr: regexp_replace('abc', 'z', 'a')
+                      type: string
+                      expr: regexp_replace('abbbb', 'bb', 'b')
+                      type: string
+                      expr: regexp_replace('hadoop', '(.)[a-z]*', '$1ive')
+                      type: string
+                      expr: regexp_replace('hadoopAAA', 'A.*', '')
+                      type: string
+                      expr: regexp_replace('abc', '', 'A')
+                      type: string
+                      expr: ('abc' regexp '')
+                      type: boolean
+                File Output Operator
+                  compressed: false
+                  GlobalTableId: 1
+                  table:
+                      input format: org.apache.hadoop.mapred.TextInputFormat
+                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                      name: dest1
 
   Stage: Stage-0
     Move Operator
@@ -104,5 +100,5 @@
 Output: default/dest1
 query: SELECT dest1.* FROM dest1
 Input: default/dest1
-Output: /data/users/zshao/tools/495-trunk-apache-hive/ql/../build/ql/tmp/572903932/16944307.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/516885779/10000
 true	false	true	true	true	false	false	false	true	true	false	true	false	acc	abc	abb	hive	hadoop	AaAbAcA	false

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf3.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf3.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf3.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf3.q.out Mon Jun 15 05:43:22 2009
@@ -27,36 +27,36 @@
                   sort order: 
                   tag: -1
                   value expressions:
-                        expr: 0
+                        expr: _col0
                         type: bigint
-                        expr: 1
+                        expr: _col1
                         type: double
-                        expr: 2
+                        expr: _col2
                         type: string
-                        expr: 3
+                        expr: _col3
                         type: int
-                        expr: 4
+                        expr: _col4
                         type: int
       Reduce Operator Tree:
         Group By Operator
           aggregations:
-                expr: count(VALUE.0)
-                expr: sum(VALUE.1)
-                expr: avg(VALUE.2)
-                expr: min(VALUE.3)
-                expr: max(VALUE.4)
+                expr: count(VALUE._col0)
+                expr: sum(VALUE._col1)
+                expr: avg(VALUE._col2)
+                expr: min(VALUE._col3)
+                expr: max(VALUE._col4)
           mode: mergepartial
           Select Operator
             expressions:
-                  expr: 0
+                  expr: _col0
                   type: bigint
-                  expr: 1
+                  expr: _col1
                   type: double
-                  expr: 2
+                  expr: _col2
                   type: double
-                  expr: 3
+                  expr: _col3
                   type: int
-                  expr: 4
+                  expr: _col4
                   type: int
             File Output Operator
               compressed: false
@@ -84,5 +84,5 @@
 Output: default/dest1
 query: SELECT dest1.* FROM dest1
 Input: default/dest1
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/441448894/123433014.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1041569072/10000
 0	NULL	NULL	NULL	NULL

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf4.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf4.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf4.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf4.q.out Mon Jun 15 05:43:22 2009
@@ -17,52 +17,51 @@
       Alias -> Map Operator Tree:
         dest1 
             Select Operator
-              Select Operator
-                expressions:
-                      expr: round(1.0)
-                      type: bigint
-                      expr: round(1.5)
-                      type: bigint
-                      expr: round(- 1.5)
-                      type: bigint
-                      expr: floor(1.0)
-                      type: bigint
-                      expr: floor(1.5)
-                      type: bigint
-                      expr: floor(- 1.5)
-                      type: bigint
-                      expr: sqrt(1.0)
-                      type: double
-                      expr: sqrt(- 1.0)
-                      type: double
-                      expr: sqrt(0.0)
-                      type: double
-                      expr: ceiling(1.0)
-                      type: bigint
-                      expr: ceiling(1.5)
-                      type: bigint
-                      expr: ceiling(- 1.5)
-                      type: bigint
-                      expr: ceiling(1.0)
-                      type: bigint
-                      expr: rand(UDFToLong(3))
-                      type: double
-                      expr: 3
-                      type: int
-                      expr: - 3
-                      type: int
-                      expr: (1 + 2)
-                      type: int
-                      expr: (1 + - 2)
-                      type: int
-                      expr: ~ 1
-                      type: int
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              expressions:
+                    expr: round(1.0)
+                    type: bigint
+                    expr: round(1.5)
+                    type: bigint
+                    expr: round(- 1.5)
+                    type: bigint
+                    expr: floor(1.0)
+                    type: bigint
+                    expr: floor(1.5)
+                    type: bigint
+                    expr: floor(- 1.5)
+                    type: bigint
+                    expr: sqrt(1.0)
+                    type: double
+                    expr: sqrt(- 1.0)
+                    type: double
+                    expr: sqrt(0.0)
+                    type: double
+                    expr: ceiling(1.0)
+                    type: bigint
+                    expr: ceiling(1.5)
+                    type: bigint
+                    expr: ceiling(- 1.5)
+                    type: bigint
+                    expr: ceiling(1.0)
+                    type: bigint
+                    expr: rand(UDFToLong(3))
+                    type: double
+                    expr: 3
+                    type: int
+                    expr: - 3
+                    type: int
+                    expr: (1 + 2)
+                    type: int
+                    expr: (1 + - 2)
+                    type: int
+                    expr: ~ 1
+                    type: int
+              File Output Operator
+                compressed: false
+                GlobalTableId: 0
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
   Stage: Stage-0
     Fetch Operator
@@ -71,5 +70,5 @@
 
 query: SELECT round(1.0), round(1.5), round(-1.5), floor(1.0), floor(1.5), floor(-1.5), sqrt(1.0), sqrt(-1.0), sqrt(0.0), ceil(1.0), ceil(1.5), ceil(-1.5), ceiling(1.0), rand(3), +3, -3, 1++2, 1+-2, ~1 FROM dest1
 Input: default/dest1
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/29787657/1016730635.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1468603599/10000
 1	2	-2	1	1	-2	1.0	NULL	0.0	1	2	-1	1	0.731057369148862	3	-3	3	-1	-2

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf5.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf5.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf5.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf5.q.out Mon Jun 15 05:43:22 2009
@@ -17,30 +17,29 @@
       Alias -> Map Operator Tree:
         dest1 
             Select Operator
-              Select Operator
-                expressions:
-                      expr: from_unixtime(1226446340)
-                      type: string
-                      expr: to_date(from_unixtime(1226446340))
-                      type: string
-                      expr: day('2008-11-01')
-                      type: int
-                      expr: month('2008-11-01')
-                      type: int
-                      expr: year('2008-11-01')
-                      type: int
-                      expr: day('2008-11-01 15:32:20')
-                      type: int
-                      expr: month('2008-11-01 15:32:20')
-                      type: int
-                      expr: year('2008-11-01 15:32:20')
-                      type: int
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              expressions:
+                    expr: from_unixtime(1226446340)
+                    type: string
+                    expr: to_date(from_unixtime(1226446340))
+                    type: string
+                    expr: day('2008-11-01')
+                    type: int
+                    expr: month('2008-11-01')
+                    type: int
+                    expr: year('2008-11-01')
+                    type: int
+                    expr: day('2008-11-01 15:32:20')
+                    type: int
+                    expr: month('2008-11-01 15:32:20')
+                    type: int
+                    expr: year('2008-11-01 15:32:20')
+                    type: int
+              File Output Operator
+                compressed: false
+                GlobalTableId: 0
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
   Stage: Stage-0
     Fetch Operator
@@ -49,5 +48,5 @@
 
 query: SELECT from_unixtime(1226446340), to_date(from_unixtime(1226446340)), day('2008-11-01'), month('2008-11-01'), year('2008-11-01'), day('2008-11-01 15:32:20'), month('2008-11-01 15:32:20'), year('2008-11-01 15:32:20') FROM dest1
 Input: default/dest1
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/396140468/103691975.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/467850962/10000
 2008-11-11 15:32:20	2008-11-11	1	11	2008	1	11	2008

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf6.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf6.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf6.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf6.q.out Mon Jun 15 05:43:22 2009
@@ -17,16 +17,15 @@
       Alias -> Map Operator Tree:
         dest1 
             Select Operator
-              Select Operator
-                expressions:
-                      expr: if true, 1, 2
-                      type: int
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              expressions:
+                    expr: if true, 1, 2
+                    type: int
+              File Output Operator
+                compressed: false
+                GlobalTableId: 0
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
   Stage: Stage-0
     Fetch Operator
@@ -35,7 +34,7 @@
 
 query: SELECT IF(TRUE, 1, 2) FROM dest1
 Input: default/dest1
-Output: /data/users/zshao/tools/495-trunk-apache-hive/ql/../build/ql/tmp/185617439/895542101.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/923424840/10000
 1
 query: EXPLAIN
 SELECT IF(TRUE, 1, 2), IF(FALSE, 1, 2), IF(NULL, 1, 2), IF(TRUE, "a", "b"),
@@ -56,36 +55,35 @@
       Alias -> Map Operator Tree:
         dest1 
             Select Operator
-              Select Operator
-                expressions:
-                      expr: if true, 1, 2
-                      type: int
-                      expr: if false, 1, 2
-                      type: int
-                      expr: if null, 1, 2
-                      type: int
-                      expr: if true, 'a', 'b'
-                      type: string
-                      expr: if true, 0.1, 0.2
-                      type: double
-                      expr: if false, UDFToLong(1), UDFToLong(2)
-                      type: bigint
-                      expr: if false, UDFToInteger(UDFToByte(127)), 128
-                      type: int
-                      expr: if false, UDFToShort(UDFToByte(127)), UDFToShort(128)
-                      type: smallint
-                      expr: 128
-                      type: int
-                      expr: 1.0
-                      type: double
-                      expr: '128'
-                      type: string
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              expressions:
+                    expr: if true, 1, 2
+                    type: int
+                    expr: if false, 1, 2
+                    type: int
+                    expr: if null, 1, 2
+                    type: int
+                    expr: if true, 'a', 'b'
+                    type: string
+                    expr: if true, 0.1, 0.2
+                    type: double
+                    expr: if false, UDFToLong(1), UDFToLong(2)
+                    type: bigint
+                    expr: if false, UDFToInteger(UDFToByte(127)), 128
+                    type: int
+                    expr: if false, UDFToShort(UDFToByte(127)), UDFToShort(128)
+                    type: smallint
+                    expr: 128
+                    type: int
+                    expr: 1.0
+                    type: double
+                    expr: '128'
+                    type: string
+              File Output Operator
+                compressed: false
+                GlobalTableId: 0
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
   Stage: Stage-0
     Fetch Operator
@@ -98,5 +96,5 @@
        CAST(128 AS SMALLINT)), CAST(128 AS INT), CAST(1.0 AS DOUBLE),
        CAST('128' AS STRING) FROM dest1
 Input: default/dest1
-Output: /data/users/zshao/tools/495-trunk-apache-hive/ql/../build/ql/tmp/238198611/144848990.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1340342481/10000
 1	2	2	a	0.1	2	128	128	128	1.0	128

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf7.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf7.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf7.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf7.q.out Mon Jun 15 05:43:22 2009
@@ -22,64 +22,63 @@
       Alias -> Map Operator Tree:
         dest1 
             Select Operator
-              Select Operator
-                expressions:
-                      expr: round(ln(3.0), 12)
-                      type: double
-                      expr: ln(0.0)
-                      type: double
-                      expr: ln(UDFToDouble(- 1))
-                      type: double
-                      expr: round(log(3.0), 12)
-                      type: double
-                      expr: log(0.0)
-                      type: double
-                      expr: log(UDFToDouble(- 1))
-                      type: double
-                      expr: round(log2(3.0), 12)
-                      type: double
-                      expr: log2(0.0)
-                      type: double
-                      expr: log2(UDFToDouble(- 1))
-                      type: double
-                      expr: round(log10(3.0), 12)
-                      type: double
-                      expr: log10(0.0)
-                      type: double
-                      expr: log10(UDFToDouble(- 1))
-                      type: double
-                      expr: round(log(UDFToDouble(2), 3.0), 12)
-                      type: double
-                      expr: log(UDFToDouble(2), 0.0)
-                      type: double
-                      expr: log(UDFToDouble(2), UDFToDouble(- 1))
-                      type: double
-                      expr: log(0.5, UDFToDouble(2))
-                      type: double
-                      expr: log(UDFToDouble(2), 0.5)
-                      type: double
-                      expr: round(exp(2.0), 12)
-                      type: double
-                      expr: pow(UDFToDouble(2), UDFToDouble(3))
-                      type: double
-                      expr: pow(UDFToDouble(2), UDFToDouble(3))
-                      type: double
-                      expr: pow(UDFToDouble(2), UDFToDouble(- 3))
-                      type: double
-                      expr: pow(0.5, UDFToDouble(- 3))
-                      type: double
-                      expr: pow(UDFToDouble(4), 0.5)
-                      type: double
-                      expr: pow(UDFToDouble(- 1), 0.5)
-                      type: double
-                      expr: pow(UDFToDouble(- 1), UDFToDouble(2))
-                      type: double
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              expressions:
+                    expr: round(ln(3.0), 12)
+                    type: double
+                    expr: ln(0.0)
+                    type: double
+                    expr: ln(UDFToDouble(- 1))
+                    type: double
+                    expr: round(log(3.0), 12)
+                    type: double
+                    expr: log(0.0)
+                    type: double
+                    expr: log(UDFToDouble(- 1))
+                    type: double
+                    expr: round(log2(3.0), 12)
+                    type: double
+                    expr: log2(0.0)
+                    type: double
+                    expr: log2(UDFToDouble(- 1))
+                    type: double
+                    expr: round(log10(3.0), 12)
+                    type: double
+                    expr: log10(0.0)
+                    type: double
+                    expr: log10(UDFToDouble(- 1))
+                    type: double
+                    expr: round(log(UDFToDouble(2), 3.0), 12)
+                    type: double
+                    expr: log(UDFToDouble(2), 0.0)
+                    type: double
+                    expr: log(UDFToDouble(2), UDFToDouble(- 1))
+                    type: double
+                    expr: log(0.5, UDFToDouble(2))
+                    type: double
+                    expr: log(UDFToDouble(2), 0.5)
+                    type: double
+                    expr: round(exp(2.0), 12)
+                    type: double
+                    expr: pow(UDFToDouble(2), UDFToDouble(3))
+                    type: double
+                    expr: pow(UDFToDouble(2), UDFToDouble(3))
+                    type: double
+                    expr: pow(UDFToDouble(2), UDFToDouble(- 3))
+                    type: double
+                    expr: pow(0.5, UDFToDouble(- 3))
+                    type: double
+                    expr: pow(UDFToDouble(4), 0.5)
+                    type: double
+                    expr: pow(UDFToDouble(- 1), 0.5)
+                    type: double
+                    expr: pow(UDFToDouble(- 1), UDFToDouble(2))
+                    type: double
+              File Output Operator
+                compressed: false
+                GlobalTableId: 0
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
   Stage: Stage-0
     Fetch Operator
@@ -93,5 +92,5 @@
        POW(2,3), POWER(2,3), POWER(2,-3), POWER(0.5, -3), POWER(4, 0.5),
        POWER(-1, 0.5), POWER(-1, 2) FROM dest1
 Input: default/dest1
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/323872951/193616753.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/987082819/10000
 1.098612288668	NULL	NULL	1.098612288668	NULL	NULL	1.584962500721	NULL	NULL	0.47712125472	NULL	NULL	1.584962500721	NULL	NULL	NULL	-1.0	7.389056098931	8.0	8.0	0.125	8.0	2.0	NaN	1.0

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf8.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf8.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf8.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf8.q.out Mon Jun 15 05:43:22 2009
@@ -19,36 +19,40 @@
     Map Reduce
       Alias -> Map Operator Tree:
         dest1 
-            Group By Operator
-              aggregations:
-                    expr: avg(UDFToDouble(c1))
-                    expr: sum(UDFToDouble(c1))
-                    expr: count(c1)
-              mode: hash
-              Reduce Output Operator
-                sort order: 
-                tag: -1
-                value expressions:
-                      expr: 0
-                      type: string
-                      expr: 1
-                      type: double
-                      expr: 2
-                      type: bigint
+            Select Operator
+              expressions:
+                    expr: c1
+                    type: string
+              Group By Operator
+                aggregations:
+                      expr: avg(UDFToDouble(c1))
+                      expr: sum(UDFToDouble(c1))
+                      expr: count(c1)
+                mode: hash
+                Reduce Output Operator
+                  sort order: 
+                  tag: -1
+                  value expressions:
+                        expr: _col0
+                        type: string
+                        expr: _col1
+                        type: double
+                        expr: _col2
+                        type: bigint
       Reduce Operator Tree:
         Group By Operator
           aggregations:
-                expr: avg(VALUE.0)
-                expr: sum(VALUE.1)
-                expr: count(VALUE.2)
+                expr: avg(VALUE._col0)
+                expr: sum(VALUE._col1)
+                expr: count(VALUE._col2)
           mode: mergepartial
           Select Operator
             expressions:
-                  expr: 0
+                  expr: _col0
                   type: double
-                  expr: 1
+                  expr: _col1
                   type: double
-                  expr: 2
+                  expr: _col2
                   type: bigint
             File Output Operator
               compressed: false
@@ -64,5 +68,5 @@
 
 query: SELECT avg(c1), sum(c1), count(c1) FROM dest1
 Input: default/dest1
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/148346145/763466676.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1344361217/10000
 1.0	1.0	1

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf9.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf9.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf9.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf9.q.out Mon Jun 15 05:43:22 2009
@@ -20,54 +20,50 @@
     Map Reduce
       Alias -> Map Operator Tree:
         src 
-            Select Operator
-              expressions:
-                    expr: key
-                    type: string
-              Filter Operator
-                predicate:
-                    expr: (UDFToDouble(0) = UDFToDouble(86))
-                    type: boolean
-                Select Operator
-                  expressions:
-                        expr: datediff('2008-12-31', '2009-01-01')
-                        type: int
-                        expr: datediff('2008-03-01', '2008-02-28')
-                        type: int
-                        expr: datediff('2007-03-01', '2007-01-28')
-                        type: int
-                        expr: datediff('2008-03-01 23:59:59', '2008-03-02 00:00:00')
-                        type: int
-                        expr: date_add('2008-12-31', 1)
-                        type: string
-                        expr: date_add('2008-12-31', 365)
-                        type: string
-                        expr: date_add('2008-02-28', 2)
-                        type: string
-                        expr: date_add('2009-02-28', 2)
-                        type: string
-                        expr: date_add('2007-02-28', 365)
-                        type: string
-                        expr: date_add('2007-02-28 23:59:59', 730)
-                        type: string
-                        expr: date_sub('2009-01-01', 1)
-                        type: string
-                        expr: date_sub('2009-01-01', 365)
-                        type: string
-                        expr: date_sub('2008-02-28', 2)
-                        type: string
-                        expr: date_sub('2009-02-28', 2)
-                        type: string
-                        expr: date_sub('2007-02-28', 365)
-                        type: string
-                        expr: date_sub('2007-02-28 01:12:34', 730)
-                        type: string
-                  File Output Operator
-                    compressed: false
-                    GlobalTableId: 0
-                    table:
-                        input format: org.apache.hadoop.mapred.TextInputFormat
-                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+            Filter Operator
+              predicate:
+                  expr: (UDFToDouble(key) = UDFToDouble(86))
+                  type: boolean
+              Select Operator
+                expressions:
+                      expr: datediff('2008-12-31', '2009-01-01')
+                      type: int
+                      expr: datediff('2008-03-01', '2008-02-28')
+                      type: int
+                      expr: datediff('2007-03-01', '2007-01-28')
+                      type: int
+                      expr: datediff('2008-03-01 23:59:59', '2008-03-02 00:00:00')
+                      type: int
+                      expr: date_add('2008-12-31', 1)
+                      type: string
+                      expr: date_add('2008-12-31', 365)
+                      type: string
+                      expr: date_add('2008-02-28', 2)
+                      type: string
+                      expr: date_add('2009-02-28', 2)
+                      type: string
+                      expr: date_add('2007-02-28', 365)
+                      type: string
+                      expr: date_add('2007-02-28 23:59:59', 730)
+                      type: string
+                      expr: date_sub('2009-01-01', 1)
+                      type: string
+                      expr: date_sub('2009-01-01', 365)
+                      type: string
+                      expr: date_sub('2008-02-28', 2)
+                      type: string
+                      expr: date_sub('2009-02-28', 2)
+                      type: string
+                      expr: date_sub('2007-02-28', 365)
+                      type: string
+                      expr: date_sub('2007-02-28 01:12:34', 730)
+                      type: string
+                File Output Operator
+                  compressed: false
+                  GlobalTableId: 0
+                  table:
+                      input format: org.apache.hadoop.mapred.TextInputFormat
+                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
   Stage: Stage-0
     Fetch Operator
@@ -84,5 +80,5 @@
        DATE_SUB('2007-02-28', 365), DATE_SUB('2007-02-28 01:12:34', 730)
        FROM src WHERE src.key = 86
 Input: default/src
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/804319999/376709074.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/2099135137/10000
 -1	2	32	-1	2009-01-01	2009-12-31	2008-03-01	2009-03-02	2008-02-28	2009-02-27	2008-12-31	2008-01-02	2008-02-28	2009-02-27	2006-02-28	2005-02-28

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_10_trims.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_10_trims.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_10_trims.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_10_trims.q.out Mon Jun 15 05:43:22 2009
@@ -16,26 +16,22 @@
     Map Reduce
       Alias -> Map Operator Tree:
         src 
-            Select Operator
-              expressions:
-                    expr: key
-                    type: string
-              Filter Operator
-                predicate:
-                    expr: (UDFToDouble(0) = UDFToDouble(86))
-                    type: boolean
-                Select Operator
-                  expressions:
-                        expr: trim(trim(trim(trim(trim(trim(trim(trim(trim(trim('  abc  '))))))))))
-                        type: string
-                  File Output Operator
-                    compressed: false
-                    GlobalTableId: 1
-                    table:
-                        input format: org.apache.hadoop.mapred.TextInputFormat
-                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                        name: dest1
+            Filter Operator
+              predicate:
+                  expr: (UDFToDouble(key) = UDFToDouble(86))
+                  type: boolean
+              Select Operator
+                expressions:
+                      expr: trim(trim(trim(trim(trim(trim(trim(trim(trim(trim('  abc  '))))))))))
+                      type: string
+                File Output Operator
+                  compressed: false
+                  GlobalTableId: 1
+                  table:
+                      input format: org.apache.hadoop.mapred.TextInputFormat
+                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                      name: dest1
 
   Stage: Stage-0
     Move Operator

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case.q.out Mon Jun 15 05:43:22 2009
@@ -38,27 +38,26 @@
       Alias -> Map Operator Tree:
         src 
             Select Operator
-              Select Operator
-                expressions:
-                      expr: CASE (1) WHEN (1) THEN (2) WHEN (3) THEN (4) ELSE (5) END
-                      type: int
-                      expr: CASE (2) WHEN (1) THEN (2) ELSE (5) END
-                      type: int
-                      expr: CASE (14) WHEN (12) THEN (13) WHEN (14) THEN (15) END
-                      type: int
-                      expr: CASE (16) WHEN (12) THEN (13) WHEN (14) THEN (15) END
-                      type: int
-                      expr: CASE (17) WHEN (18) THEN (null) WHEN (17) THEN (20) END
-                      type: int
-                      expr: CASE (21) WHEN (22) THEN (23) WHEN (21) THEN (24) END
-                      type: int
-                Limit
-                  File Output Operator
-                    compressed: false
-                    GlobalTableId: 0
-                    table:
-                        input format: org.apache.hadoop.mapred.TextInputFormat
-                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              expressions:
+                    expr: CASE (1) WHEN (1) THEN (2) WHEN (3) THEN (4) ELSE (5) END
+                    type: int
+                    expr: CASE (2) WHEN (1) THEN (2) ELSE (5) END
+                    type: int
+                    expr: CASE (14) WHEN (12) THEN (13) WHEN (14) THEN (15) END
+                    type: int
+                    expr: CASE (16) WHEN (12) THEN (13) WHEN (14) THEN (15) END
+                    type: int
+                    expr: CASE (17) WHEN (18) THEN (null) WHEN (17) THEN (20) END
+                    type: int
+                    expr: CASE (21) WHEN (22) THEN (23) WHEN (21) THEN (24) END
+                    type: int
+              Limit
+                File Output Operator
+                  compressed: false
+                  GlobalTableId: 0
+                  table:
+                      input format: org.apache.hadoop.mapred.TextInputFormat
+                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
   Stage: Stage-0
     Fetch Operator
@@ -92,5 +91,5 @@
        END
 FROM src LIMIT 1
 Input: default/src
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/116746963/155290781.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1131510224/10000
 2	5	15	NULL	20	24

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case_column_pruning.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case_column_pruning.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case_column_pruning.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case_column_pruning.q.out Mon Jun 15 05:43:22 2009
@@ -20,49 +20,45 @@
     Map Reduce
       Alias -> Map Operator Tree:
         b 
-            Select Operator
-              expressions:
+            Reduce Output Operator
+              key expressions:
+                    expr: key
+                    type: string
+              sort order: +
+              Map-reduce partition columns:
                     expr: key
                     type: string
-              Reduce Output Operator
-                key expressions:
-                      expr: 0
-                      type: string
-                sort order: +
-                Map-reduce partition columns:
-                      expr: 0
-                      type: string
-                tag: 1
-                value expressions:
-                      expr: 0
-                      type: string
+              tag: 1
+              value expressions:
+                    expr: key
+                    type: string
+                    expr: value
+                    type: string
         a 
-            Select Operator
-              expressions:
+            Reduce Output Operator
+              key expressions:
+                    expr: key
+                    type: string
+              sort order: +
+              Map-reduce partition columns:
                     expr: key
                     type: string
-              Reduce Output Operator
-                key expressions:
-                      expr: 0
-                      type: string
-                sort order: +
-                Map-reduce partition columns:
-                      expr: 0
-                      type: string
-                tag: 0
-                value expressions:
-                      expr: 0
-                      type: string
+              tag: 0
+              value expressions:
+                    expr: key
+                    type: string
+                    expr: value
+                    type: string
       Reduce Operator Tree:
         Join Operator
           condition map:
                Inner Join 0 to 1
           condition expressions:
-            0 {VALUE.0}
-            1 {VALUE.0}
+            0 {VALUE._col0} {VALUE._col1}
+            1 {VALUE._col0} {VALUE._col1}
           Select Operator
             expressions:
-                  expr: CASE (0) WHEN ('1') THEN (2) WHEN ('3') THEN (4) ELSE (5) END
+                  expr: CASE (_col0) WHEN ('1') THEN (2) WHEN ('3') THEN (4) ELSE (5) END
                   type: int
             File Output Operator
               compressed: false
@@ -75,15 +71,15 @@
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/data/users/njain/hive2/hive2/build/ql/tmp/982397491/10004 
+        file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1081703636/10002 
             Reduce Output Operator
               key expressions:
-                    expr: 0
+                    expr: _col0
                     type: int
               sort order: +
               tag: -1
               value expressions:
-                    expr: 0
+                    expr: _col0
                     type: int
       Reduce Operator Tree:
         Extract
@@ -109,7 +105,7 @@
 ON a.key = b.key
 ORDER BY key LIMIT 10
 Input: default/src
-Output: file:/data/users/njain/hive2/hive2/build/ql/tmp/163558385/10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/10034618/10000
 5
 5
 5

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case_thrift.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case_thrift.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case_thrift.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case_thrift.q.out Mon Jun 15 05:43:22 2009
@@ -29,25 +29,19 @@
         src_thrift 
             Select Operator
               expressions:
-                    expr: lint
-                    type: array<int>
-                    expr: lstring
-                    type: array<string>
-              Select Operator
-                expressions:
-                      expr: CASE (0[0]) WHEN (0) THEN ((0[0] + 1)) WHEN (1) THEN ((0[0] + 2)) WHEN (2) THEN (100) ELSE (5) END
-                      type: int
-                      expr: CASE (1[0]) WHEN ('0') THEN ('zero') WHEN ('10') THEN (concat(1[0], ' is ten')) ELSE ('default') END
-                      type: string
-                      expr: CASE (1[0]) WHEN ('0') THEN (1) ELSE (null) END[0]
-                      type: string
-                Limit
-                  File Output Operator
-                    compressed: false
-                    GlobalTableId: 0
-                    table:
-                        input format: org.apache.hadoop.mapred.TextInputFormat
-                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                    expr: CASE (lint[0]) WHEN (0) THEN ((lint[0] + 1)) WHEN (1) THEN ((lint[0] + 2)) WHEN (2) THEN (100) ELSE (5) END
+                    type: int
+                    expr: CASE (lstring[0]) WHEN ('0') THEN ('zero') WHEN ('10') THEN (concat(lstring[0], ' is ten')) ELSE ('default') END
+                    type: string
+                    expr: CASE (lstring[0]) WHEN ('0') THEN (lstring) ELSE (null) END[0]
+                    type: string
+              Limit
+                File Output Operator
+                  compressed: false
+                  GlobalTableId: 0
+                  table:
+                      input format: org.apache.hadoop.mapred.TextInputFormat
+                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
   Stage: Stage-0
     Fetch Operator
@@ -71,7 +65,7 @@
        END)[0]
 FROM src_thrift LIMIT 3
 Input: default/src_thrift
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/1401002108/1327666619.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1872084071/10000
 1	zero	0
 3	10 is ten	NULL
 100	default	NULL

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_coalesce.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_coalesce.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_coalesce.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_coalesce.q.out Mon Jun 15 05:43:22 2009
@@ -31,51 +31,50 @@
       Alias -> Map Operator Tree:
         src 
             Select Operator
-              Select Operator
-                expressions:
-                      expr: COALESCE(1)
-                      type: int
-                      expr: COALESCE(1,2)
-                      type: int
-                      expr: COALESCE(null,2)
-                      type: int
-                      expr: COALESCE(1,null)
-                      type: int
-                      expr: COALESCE(null,null,3)
-                      type: int
-                      expr: COALESCE(4,null,null,null)
-                      type: int
-                      expr: COALESCE('1')
-                      type: string
-                      expr: COALESCE('1','2')
-                      type: string
-                      expr: COALESCE(null,'2')
-                      type: string
-                      expr: COALESCE('1',null)
-                      type: string
-                      expr: COALESCE(null,null,'3')
-                      type: string
-                      expr: COALESCE('4',null,null,null)
-                      type: string
-                      expr: COALESCE(1.0)
-                      type: double
-                      expr: COALESCE(1.0,2.0)
-                      type: double
-                      expr: COALESCE(null,2.0)
-                      type: double
-                      expr: COALESCE(null,2.0,3.0)
-                      type: double
-                      expr: COALESCE(2.0,null,3.0)
-                      type: double
-                      expr: COALESCE(if true, null, 0,null)
-                      type: int
-                Limit
-                  File Output Operator
-                    compressed: false
-                    GlobalTableId: 0
-                    table:
-                        input format: org.apache.hadoop.mapred.TextInputFormat
-                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              expressions:
+                    expr: COALESCE(1)
+                    type: int
+                    expr: COALESCE(1,2)
+                    type: int
+                    expr: COALESCE(null,2)
+                    type: int
+                    expr: COALESCE(1,null)
+                    type: int
+                    expr: COALESCE(null,null,3)
+                    type: int
+                    expr: COALESCE(4,null,null,null)
+                    type: int
+                    expr: COALESCE('1')
+                    type: string
+                    expr: COALESCE('1','2')
+                    type: string
+                    expr: COALESCE(null,'2')
+                    type: string
+                    expr: COALESCE('1',null)
+                    type: string
+                    expr: COALESCE(null,null,'3')
+                    type: string
+                    expr: COALESCE('4',null,null,null)
+                    type: string
+                    expr: COALESCE(1.0)
+                    type: double
+                    expr: COALESCE(1.0,2.0)
+                    type: double
+                    expr: COALESCE(null,2.0)
+                    type: double
+                    expr: COALESCE(null,2.0,3.0)
+                    type: double
+                    expr: COALESCE(2.0,null,3.0)
+                    type: double
+                    expr: COALESCE(if true, null, 0,null)
+                    type: int
+              Limit
+                File Output Operator
+                  compressed: false
+                  GlobalTableId: 0
+                  table:
+                      input format: org.apache.hadoop.mapred.TextInputFormat
+                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
   Stage: Stage-0
     Fetch Operator
@@ -102,7 +101,7 @@
        COALESCE(IF(TRUE, NULL, 0), NULL)
 FROM src LIMIT 1
 Input: default/src
-Output: file:/data/users/zshao/tools/495-trunk-apache-hive/build/ql/tmp/64965527/10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1424226650/10000
 1	1	2	1	3	4	1	1	2	1	3	4	1.0	1.0	2.0	2.0	2.0	NULL
 query: EXPLAIN
 SELECT COALESCE(src_thrift.lint[1], 999),
@@ -123,26 +122,18 @@
         src_thrift 
             Select Operator
               expressions:
-                    expr: lint
-                    type: array<int>
-                    expr: lintstring
-                    type: array<struct<myint:int,mystring:string,underscore_int:int>>
-                    expr: mstringstring
-                    type: map<string,string>
-              Select Operator
-                expressions:
-                      expr: COALESCE(0[1],999)
-                      type: int
-                      expr: COALESCE(1[0].mystring,'999')
-                      type: string
-                      expr: COALESCE(2['key_2'],'999')
-                      type: string
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                    expr: COALESCE(lint[1],999)
+                    type: int
+                    expr: COALESCE(lintstring[0].mystring,'999')
+                    type: string
+                    expr: COALESCE(mstringstring['key_2'],'999')
+                    type: string
+              File Output Operator
+                compressed: false
+                GlobalTableId: 0
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
   Stage: Stage-0
     Fetch Operator
@@ -154,7 +145,7 @@
        COALESCE(src_thrift.mstringstring['key_2'], '999')
 FROM src_thrift
 Input: default/src_thrift
-Output: file:/data/users/zshao/tools/495-trunk-apache-hive/build/ql/tmp/2056249139/10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/300664928/10000
 0	0	999
 2	1	999
 4	8	value_2

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_hash.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_hash.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_hash.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_hash.q.out Mon Jun 15 05:43:22 2009
@@ -18,37 +18,36 @@
       Alias -> Map Operator Tree:
         src 
             Select Operator
-              Select Operator
-                expressions:
-                      expr: hash(UDFToByte(1))
-                      type: int
-                      expr: hash(UDFToShort(2))
-                      type: int
-                      expr: hash(3)
-                      type: int
-                      expr: hash(UDFToLong('123456789012'))
-                      type: int
-                      expr: hash(UDFToFloat(1.25))
-                      type: int
-                      expr: hash(16.0)
-                      type: int
-                      expr: hash('400')
-                      type: int
-                      expr: hash('abc')
-                      type: int
-                      expr: hash(true)
-                      type: int
-                      expr: hash(false)
-                      type: int
-                      expr: hash(1,2,3)
-                      type: int
-                Limit
-                  File Output Operator
-                    compressed: false
-                    GlobalTableId: 0
-                    table:
-                        input format: org.apache.hadoop.mapred.TextInputFormat
-                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              expressions:
+                    expr: hash(UDFToByte(1))
+                    type: int
+                    expr: hash(UDFToShort(2))
+                    type: int
+                    expr: hash(3)
+                    type: int
+                    expr: hash(UDFToLong('123456789012'))
+                    type: int
+                    expr: hash(UDFToFloat(1.25))
+                    type: int
+                    expr: hash(16.0)
+                    type: int
+                    expr: hash('400')
+                    type: int
+                    expr: hash('abc')
+                    type: int
+                    expr: hash(true)
+                    type: int
+                    expr: hash(false)
+                    type: int
+                    expr: hash(1,2,3)
+                    type: int
+              Limit
+                File Output Operator
+                  compressed: false
+                  GlobalTableId: 0
+                  table:
+                      input format: org.apache.hadoop.mapred.TextInputFormat
+                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
   Stage: Stage-0
     Fetch Operator
@@ -62,5 +61,5 @@
        hash(1, 2, 3)
 FROM src LIMIT 1
 Input: default/src
-Output: /data/users/zshao/tools/495-trunk-apache-hive/ql/../build/ql/tmp/347375001/1065831137.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1354544735/10000
 1	2	3	-1097262584	1067450368	1076887552	51508	96354	1	0	1026

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_length.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_length.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_length.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_length.q.out Mon Jun 15 05:43:22 2009
@@ -15,20 +15,16 @@
         src 
             Select Operator
               expressions:
-                    expr: value
-                    type: string
-              Select Operator
-                expressions:
-                      expr: length(0)
-                      type: int
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 1
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                      name: dest1
+                    expr: length(value)
+                    type: int
+              File Output Operator
+                compressed: false
+                GlobalTableId: 1
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                    name: dest1
 
   Stage: Stage-0
     Move Operator
@@ -46,7 +42,7 @@
 Output: default/dest1
 query: SELECT dest1.* FROM dest1
 Input: default/dest1
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/432923810/244566699.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/423777377/10000
 7
 6
 7

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_lower.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_lower.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_lower.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_lower.q.out Mon Jun 15 05:43:22 2009
@@ -12,26 +12,22 @@
     Map Reduce
       Alias -> Map Operator Tree:
         src 
-            Select Operator
-              expressions:
-                    expr: key
-                    type: string
-              Filter Operator
-                predicate:
-                    expr: (UDFToDouble(0) = UDFToDouble(86))
-                    type: boolean
-                Select Operator
-                  expressions:
-                        expr: lower('AbC 123')
-                        type: string
-                        expr: ucase('AbC 123')
-                        type: string
-                  File Output Operator
-                    compressed: false
-                    GlobalTableId: 0
-                    table:
-                        input format: org.apache.hadoop.mapred.TextInputFormat
-                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+            Filter Operator
+              predicate:
+                  expr: (UDFToDouble(key) = UDFToDouble(86))
+                  type: boolean
+              Select Operator
+                expressions:
+                      expr: lower('AbC 123')
+                      type: string
+                      expr: ucase('AbC 123')
+                      type: string
+                File Output Operator
+                  compressed: false
+                  GlobalTableId: 0
+                  table:
+                      input format: org.apache.hadoop.mapred.TextInputFormat
+                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
   Stage: Stage-0
     Fetch Operator
@@ -40,5 +36,5 @@
 
 query: SELECT lower('AbC 123'), upper('AbC 123') FROM src WHERE key = 86
 Input: default/src
-Output: /data/users/zshao/tools/495-trunk-apache-hive/ql/../build/ql/tmp/1060690797/116583439.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1926782866/10000
 abc 123	ABC 123

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_when.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_when.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_when.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_when.q.out Mon Jun 15 05:43:22 2009
@@ -38,27 +38,26 @@
       Alias -> Map Operator Tree:
         src 
             Select Operator
-              Select Operator
-                expressions:
-                      expr: CASE WHEN ((1 = 1)) THEN (2) WHEN ((1 = 3)) THEN (4) ELSE (5) END
-                      type: int
-                      expr: CASE WHEN ((6 = 7)) THEN (8) ELSE (9) END
-                      type: int
-                      expr: CASE WHEN ((10 = 11)) THEN (12) WHEN ((13 = 13)) THEN (14) END
-                      type: int
-                      expr: CASE WHEN ((15 = 16)) THEN (17) WHEN ((18 = 19)) THEN (20) END
-                      type: int
-                      expr: CASE WHEN ((21 = 22)) THEN (null) WHEN ((23 = 23)) THEN (24) END
-                      type: int
-                      expr: CASE WHEN ((25 = 26)) THEN (27) WHEN ((28 = 28)) THEN (null) END
-                      type: int
-                Limit
-                  File Output Operator
-                    compressed: false
-                    GlobalTableId: 0
-                    table:
-                        input format: org.apache.hadoop.mapred.TextInputFormat
-                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              expressions:
+                    expr: CASE WHEN ((1 = 1)) THEN (2) WHEN ((1 = 3)) THEN (4) ELSE (5) END
+                    type: int
+                    expr: CASE WHEN ((6 = 7)) THEN (8) ELSE (9) END
+                    type: int
+                    expr: CASE WHEN ((10 = 11)) THEN (12) WHEN ((13 = 13)) THEN (14) END
+                    type: int
+                    expr: CASE WHEN ((15 = 16)) THEN (17) WHEN ((18 = 19)) THEN (20) END
+                    type: int
+                    expr: CASE WHEN ((21 = 22)) THEN (null) WHEN ((23 = 23)) THEN (24) END
+                    type: int
+                    expr: CASE WHEN ((25 = 26)) THEN (27) WHEN ((28 = 28)) THEN (null) END
+                    type: int
+              Limit
+                File Output Operator
+                  compressed: false
+                  GlobalTableId: 0
+                  table:
+                      input format: org.apache.hadoop.mapred.TextInputFormat
+                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
   Stage: Stage-0
     Fetch Operator
@@ -92,5 +91,5 @@
        END
 FROM src LIMIT 1
 Input: default/src
-Output: /data/users/athusoo/commits/hive_trunk_ws8/ql/../build/ql/tmp/262454053/648350622.10000
+Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/855982416/10000
 2	9	14	NULL	24	NULL

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/union.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/union.q.out?rev=784656&r1=784655&r2=784656&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/union.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/union.q.out Mon Jun 15 05:43:22 2009
@@ -32,9 +32,9 @@
                 Union
                   Select Operator
                     expressions:
-                          expr: 0
+                          expr: _col0
                           type: string
-                          expr: 1
+                          expr: _col1
                           type: string
                     File Output Operator
                       compressed: false
@@ -56,9 +56,9 @@
                 Union
                   Select Operator
                     expressions:
-                          expr: 0
+                          expr: _col0
                           type: string
-                          expr: 1
+                          expr: _col1
                           type: string
                     File Output Operator
                       compressed: false