You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by zs...@apache.org on 2009/06/24 06:05:05 UTC

svn commit: r787898 [5/19] - in /hadoop/hive/trunk: ./ conf/ ql/src/java/org/apache/hadoop/hive/ql/ppd/ ql/src/test/results/clientpositive/ ql/src/test/results/compiler/plan/

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out Wed Jun 24 04:05:02 2009
@@ -19,49 +19,53 @@
         s 
             Filter Operator
               predicate:
-                  expr: (((hash(key) & 2147483647) % 4) = 0)
+                  expr: ((((hash(key) & 2147483647) % 4) = 0) and (UDFToDouble(key) > UDFToDouble(100)))
                   type: boolean
               Filter Operator
                 predicate:
-                    expr: (UDFToDouble(key) > UDFToDouble(100))
+                    expr: (((hash(key) & 2147483647) % 4) = 0)
                     type: boolean
-                Select Operator
-                  expressions:
-                        expr: key
-                        type: string
-                        expr: value
-                        type: string
+                Filter Operator
+                  predicate:
+                      expr: (UDFToDouble(key) > UDFToDouble(100))
+                      type: boolean
                   Select Operator
                     expressions:
-                          expr: UDFToInteger(_col0)
-                          type: int
-                          expr: _col1
+                          expr: key
+                          type: string
+                          expr: value
                           type: string
-                    File Output Operator
-                      compressed: false
-                      GlobalTableId: 1
-                      directory: file:/data/users/njain/hive4/hive4/build/ql/tmp/862363094/10002
-                      table:
-                          input format: org.apache.hadoop.mapred.TextInputFormat
-                          output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                          properties:
-                            name dest1
-                            columns.types int:string
-                            serialization.ddl struct dest1 { i32 key, string value}
-                            serialization.format 1
-                            columns key,value
-                            bucket_count -1
-                            serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                            file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                            file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                            location file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/dest1
-                          serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                          name: dest1
+                    Select Operator
+                      expressions:
+                            expr: UDFToInteger(_col0)
+                            type: int
+                            expr: _col1
+                            type: string
+                      File Output Operator
+                        compressed: false
+                        GlobalTableId: 1
+                        directory: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1062372910/10002
+                        table:
+                            input format: org.apache.hadoop.mapred.TextInputFormat
+                            output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                            properties:
+                              name dest1
+                              columns.types int:string
+                              serialization.ddl struct dest1 { i32 key, string value}
+                              serialization.format 1
+                              columns key,value
+                              bucket_count -1
+                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                              file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                              location file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/dest1
+                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                            name: dest1
       Needs Tagging: false
       Path -> Alias:
-        file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcbucket/kv1.txt 
+        file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcbucket/kv1.txt 
       Path -> Partition:
-        file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcbucket/kv1.txt 
+        file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcbucket/kv1.txt 
           Partition
           
               input format: org.apache.hadoop.mapred.TextInputFormat
@@ -77,7 +81,7 @@
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcbucket
+                location file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcbucket
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: srcbucket
 
@@ -87,11 +91,11 @@
           Move Operator
             files:
                 hdfs directory: true
-                source: file:/data/users/njain/hive4/hive4/build/ql/tmp/862363094/10002
-                destination: file:/data/users/njain/hive4/hive4/build/ql/tmp/1120714236/10000
+                source: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1062372910/10002
+                destination: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/921828791/10000
           Map Reduce
             Alias -> Map Operator Tree:
-              file:/data/users/njain/hive4/hive4/build/ql/tmp/862363094/10002 
+              file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1062372910/10002 
                   Reduce Output Operator
                     sort order: 
                     Map-reduce partition columns:
@@ -105,9 +109,9 @@
                           type: string
             Needs Tagging: false
             Path -> Alias:
-              file:/data/users/njain/hive4/hive4/build/ql/tmp/862363094/10002 
+              file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1062372910/10002 
             Path -> Partition:
-              file:/data/users/njain/hive4/hive4/build/ql/tmp/862363094/10002 
+              file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1062372910/10002 
                 Partition
                 
                     input format: org.apache.hadoop.mapred.TextInputFormat
@@ -122,7 +126,7 @@
                       serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                       file.inputformat org.apache.hadoop.mapred.TextInputFormat
                       file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                      location file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/dest1
+                      location file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/dest1
                     serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                     name: dest1
             Reduce Operator Tree:
@@ -130,7 +134,7 @@
                 File Output Operator
                   compressed: false
                   GlobalTableId: 0
-                  directory: file:/data/users/njain/hive4/hive4/build/ql/tmp/1120714236/10000
+                  directory: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/921828791/10000
                   table:
                       input format: org.apache.hadoop.mapred.TextInputFormat
                       output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -144,7 +148,7 @@
                         serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                         file.inputformat org.apache.hadoop.mapred.TextInputFormat
                         file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                        location file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/dest1
+                        location file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/dest1
                       serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                       name: dest1
 
@@ -152,7 +156,7 @@
     Move Operator
       tables:
           replace: true
-          source: file:/data/users/njain/hive4/hive4/build/ql/tmp/1120714236/10000
+          source: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/921828791/10000
           table:
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -166,10 +170,10 @@
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/dest1
+                location file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/dest1
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: dest1
-          tmp directory: file:/data/users/njain/hive4/hive4/build/ql/tmp/1120714236/10001
+          tmp directory: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/921828791/10001
 
 
 query: INSERT OVERWRITE TABLE dest1 SELECT s.* 
@@ -179,7 +183,7 @@
 Output: default/dest1
 query: SELECT dest1.* FROM dest1
 Input: default/dest1
-Output: file:/data/users/njain/hive4/hive4/build/ql/tmp/1094962031/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1564717662/10000
 165	val_165
 484	val_484
 150	val_150

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/sample8.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/sample8.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/sample8.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/sample8.q.out Wed Jun 24 04:05:02 2009
@@ -23,43 +23,51 @@
               predicate:
                   expr: (((hash(key) & 2147483647) % 10) = 0)
                   type: boolean
-              Reduce Output Operator
-                sort order: 
-                tag: 1
-                value expressions:
-                      expr: key
-                      type: string
-                      expr: value
-                      type: string
-                      expr: ds
-                      type: string
-                      expr: hr
-                      type: string
+              Filter Operator
+                predicate:
+                    expr: (((hash(key) & 2147483647) % 10) = 0)
+                    type: boolean
+                Reduce Output Operator
+                  sort order: 
+                  tag: 1
+                  value expressions:
+                        expr: key
+                        type: string
+                        expr: value
+                        type: string
+                        expr: ds
+                        type: string
+                        expr: hr
+                        type: string
         s 
             Filter Operator
               predicate:
-                  expr: (((hash(key) & 2147483647) % 1) = 0)
+                  expr: (((((((hash(key) & 2147483647) % 1) = 0) and (ds = '2008-04-08')) and (hr = '11')) and (ds = '2008-04-08')) and (hr = '11'))
                   type: boolean
-              Reduce Output Operator
-                sort order: 
-                tag: 0
-                value expressions:
-                      expr: key
-                      type: string
-                      expr: value
-                      type: string
-                      expr: ds
-                      type: string
-                      expr: hr
-                      type: string
+              Filter Operator
+                predicate:
+                    expr: (((hash(key) & 2147483647) % 1) = 0)
+                    type: boolean
+                Reduce Output Operator
+                  sort order: 
+                  tag: 0
+                  value expressions:
+                        expr: key
+                        type: string
+                        expr: value
+                        type: string
+                        expr: ds
+                        type: string
+                        expr: hr
+                        type: string
       Needs Tagging: true
       Path -> Alias:
-        file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 
-        file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 
-        file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11 
-        file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12 
+        file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 
+        file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 
+        file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11 
+        file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12 
       Path -> Partition:
-        file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 
+        file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 
           Partition
             partition values:
               ds 2008-04-08
@@ -78,10 +86,10 @@
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart
+                location file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: srcpart
-        file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 
+        file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 
           Partition
             partition values:
               ds 2008-04-08
@@ -100,10 +108,10 @@
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart
+                location file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: srcpart
-        file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11 
+        file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11 
           Partition
             partition values:
               ds 2008-04-09
@@ -122,10 +130,10 @@
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart
+                location file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: srcpart
-        file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12 
+        file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12 
           Partition
             partition values:
               ds 2008-04-09
@@ -144,7 +152,7 @@
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart
+                location file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: srcpart
       Reduce Operator Tree:
@@ -171,7 +179,7 @@
               File Output Operator
                 compressed: false
                 GlobalTableId: 0
-                directory: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1858927587/10002
+                directory: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/709064843/10002
                 table:
                     input format: org.apache.hadoop.mapred.SequenceFileInputFormat
                     output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -184,7 +192,7 @@
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1858927587/10002 
+        file:/data/users/pchakka/workspace/oshive/build/ql/tmp/709064843/10002 
             Reduce Output Operator
               key expressions:
                     expr: _col0
@@ -209,9 +217,9 @@
                     type: string
       Needs Tagging: false
       Path -> Alias:
-        file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1858927587/10002 
+        file:/data/users/pchakka/workspace/oshive/build/ql/tmp/709064843/10002 
       Path -> Partition:
-        file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1858927587/10002 
+        file:/data/users/pchakka/workspace/oshive/build/ql/tmp/709064843/10002 
           Partition
           
               input format: org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -226,7 +234,7 @@
           File Output Operator
             compressed: false
             GlobalTableId: 0
-            directory: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1858927587/10001
+            directory: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/709064843/10001
             table:
                 input format: org.apache.hadoop.mapred.TextInputFormat
                 output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -249,7 +257,7 @@
 Input: default/srcpart/ds=2008-04-08/hr=12
 Input: default/srcpart/ds=2008-04-09/hr=11
 Input: default/srcpart/ds=2008-04-09/hr=12
-Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/587483226/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/362051028/10000
 0	val_0
 0	val_0
 0	val_0

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/subq.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/subq.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/subq.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/subq.q.out Wed Jun 24 04:05:02 2009
@@ -20,24 +20,28 @@
               predicate:
                   expr: (UDFToDouble(key) < UDFToDouble(100))
                   type: boolean
-              Select Operator
-                expressions:
-                      expr: key
-                      type: string
-                      expr: value
-                      type: string
+              Filter Operator
+                predicate:
+                    expr: (UDFToDouble(key) < UDFToDouble(100))
+                    type: boolean
                 Select Operator
                   expressions:
-                        expr: _col0
+                        expr: key
                         type: string
-                        expr: _col1
+                        expr: value
                         type: string
-                  File Output Operator
-                    compressed: false
-                    GlobalTableId: 1
-                    table:
-                        input format: org.apache.hadoop.mapred.TextInputFormat
-                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                  Select Operator
+                    expressions:
+                          expr: _col0
+                          type: string
+                          expr: _col1
+                          type: string
+                    File Output Operator
+                      compressed: false
+                      GlobalTableId: 1
+                      table:
+                          input format: org.apache.hadoop.mapred.TextInputFormat
+                          output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
   Stage: Stage-4
     Conditional Operator
@@ -45,10 +49,10 @@
           Move Operator
             files:
                 hdfs directory: true
-                destination: file:/data/users/njain/hive4/hive4/build/ql/tmp/260945618/10000
+                destination: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1386424262/10000
           Map Reduce
             Alias -> Map Operator Tree:
-              file:/data/users/njain/hive4/hive4/build/ql/tmp/1316342742/10001 
+              file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1220050565/10001 
                   Reduce Output Operator
                     sort order: 
                     Map-reduce partition columns:

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out Wed Jun 24 04:05:02 2009
@@ -14,29 +14,33 @@
     Map Reduce
       Alias -> Map Operator Tree:
         a:b 
-            Select Operator
-              expressions:
-                    expr: key
-                    type: string
-              Group By Operator
-                aggregations:
-                      expr: count(1)
-                keys:
+            Filter Operator
+              predicate:
+                  expr: (UDFToDouble(key) >= UDFToDouble(90))
+                  type: boolean
+              Select Operator
+                expressions:
                       expr: key
                       type: string
-                mode: hash
-                Reduce Output Operator
-                  key expressions:
-                        expr: _col0
-                        type: string
-                  sort order: +
-                  Map-reduce partition columns:
-                        expr: _col0
+                Group By Operator
+                  aggregations:
+                        expr: count(1)
+                  keys:
+                        expr: key
                         type: string
-                  tag: -1
-                  value expressions:
-                        expr: _col1
-                        type: bigint
+                  mode: hash
+                  Reduce Output Operator
+                    key expressions:
+                          expr: _col0
+                          type: string
+                    sort order: +
+                    Map-reduce partition columns:
+                          expr: _col0
+                          type: string
+                    tag: -1
+                    value expressions:
+                          expr: _col1
+                          type: bigint
       Reduce Operator Tree:
         Group By Operator
           aggregations:
@@ -77,7 +81,7 @@
 FROM (SELECT b.key as k, count(1) as c FROM src b GROUP BY b.key) a
 WHERE a.k >= 90
 Input: default/src
-Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/2044684588/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/259421857/10000
 100	2
 103	2
 104	2

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf1.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf1.q.out Wed Jun 24 04:05:02 2009
@@ -28,56 +28,60 @@
               predicate:
                   expr: (UDFToDouble(key) = UDFToDouble(86))
                   type: boolean
-              Select Operator
-                expressions:
-                      expr: ('a' like '%a%')
-                      type: boolean
-                      expr: ('b' like '%a%')
-                      type: boolean
-                      expr: ('ab' like '%a%')
-                      type: boolean
-                      expr: ('ab' like '%a_')
-                      type: boolean
-                      expr: ('%_' like '\%\_')
-                      type: boolean
-                      expr: ('ab' like '\%\_')
-                      type: boolean
-                      expr: ('ab' like '_a%')
-                      type: boolean
-                      expr: ('ab' like 'a')
-                      type: boolean
-                      expr: ('' regexp '.*')
-                      type: boolean
-                      expr: ('a' regexp '[ab]')
-                      type: boolean
-                      expr: ('' regexp '[ab]')
-                      type: boolean
-                      expr: ('hadoop' regexp '[a-z]*')
-                      type: boolean
-                      expr: ('hadoop' regexp 'o*')
-                      type: boolean
-                      expr: regexp_replace('abc', 'b', 'c')
-                      type: string
-                      expr: regexp_replace('abc', 'z', 'a')
-                      type: string
-                      expr: regexp_replace('abbbb', 'bb', 'b')
-                      type: string
-                      expr: regexp_replace('hadoop', '(.)[a-z]*', '$1ive')
-                      type: string
-                      expr: regexp_replace('hadoopAAA', 'A.*', '')
-                      type: string
-                      expr: regexp_replace('abc', '', 'A')
-                      type: string
-                      expr: ('abc' regexp '')
-                      type: boolean
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 1
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                      name: dest1
+              Filter Operator
+                predicate:
+                    expr: (UDFToDouble(key) = UDFToDouble(86))
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: ('a' like '%a%')
+                        type: boolean
+                        expr: ('b' like '%a%')
+                        type: boolean
+                        expr: ('ab' like '%a%')
+                        type: boolean
+                        expr: ('ab' like '%a_')
+                        type: boolean
+                        expr: ('%_' like '\%\_')
+                        type: boolean
+                        expr: ('ab' like '\%\_')
+                        type: boolean
+                        expr: ('ab' like '_a%')
+                        type: boolean
+                        expr: ('ab' like 'a')
+                        type: boolean
+                        expr: ('' regexp '.*')
+                        type: boolean
+                        expr: ('a' regexp '[ab]')
+                        type: boolean
+                        expr: ('' regexp '[ab]')
+                        type: boolean
+                        expr: ('hadoop' regexp '[a-z]*')
+                        type: boolean
+                        expr: ('hadoop' regexp 'o*')
+                        type: boolean
+                        expr: regexp_replace('abc', 'b', 'c')
+                        type: string
+                        expr: regexp_replace('abc', 'z', 'a')
+                        type: string
+                        expr: regexp_replace('abbbb', 'bb', 'b')
+                        type: string
+                        expr: regexp_replace('hadoop', '(.)[a-z]*', '$1ive')
+                        type: string
+                        expr: regexp_replace('hadoopAAA', 'A.*', '')
+                        type: string
+                        expr: regexp_replace('abc', '', 'A')
+                        type: string
+                        expr: ('abc' regexp '')
+                        type: boolean
+                  File Output Operator
+                    compressed: false
+                    GlobalTableId: 1
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                        name: dest1
 
   Stage: Stage-4
     Conditional Operator
@@ -85,10 +89,10 @@
           Move Operator
             files:
                 hdfs directory: true
-                destination: file:/data/users/njain/hive4/hive4/build/ql/tmp/1253611032/10000
+                destination: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/734075252/10000
           Map Reduce
             Alias -> Map Operator Tree:
-              file:/data/users/njain/hive4/hive4/build/ql/tmp/1620724991/10002 
+              file:/data/users/pchakka/workspace/oshive/build/ql/tmp/816992741/10002 
                   Reduce Output Operator
                     sort order: 
                     Map-reduce partition columns:
@@ -169,5 +173,5 @@
 Output: default/dest1
 query: SELECT dest1.* FROM dest1
 Input: default/dest1
-Output: file:/data/users/njain/hive4/hive4/build/ql/tmp/1990021744/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/674207509/10000
 true	false	true	true	true	false	false	false	true	true	false	true	false	acc	abc	abb	hive	hadoop	AaAbAcA	false

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf9.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf9.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf9.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf9.q.out Wed Jun 24 04:05:02 2009
@@ -24,46 +24,50 @@
               predicate:
                   expr: (UDFToDouble(key) = UDFToDouble(86))
                   type: boolean
-              Select Operator
-                expressions:
-                      expr: datediff('2008-12-31', '2009-01-01')
-                      type: int
-                      expr: datediff('2008-03-01', '2008-02-28')
-                      type: int
-                      expr: datediff('2007-03-01', '2007-01-28')
-                      type: int
-                      expr: datediff('2008-03-01 23:59:59', '2008-03-02 00:00:00')
-                      type: int
-                      expr: date_add('2008-12-31', 1)
-                      type: string
-                      expr: date_add('2008-12-31', 365)
-                      type: string
-                      expr: date_add('2008-02-28', 2)
-                      type: string
-                      expr: date_add('2009-02-28', 2)
-                      type: string
-                      expr: date_add('2007-02-28', 365)
-                      type: string
-                      expr: date_add('2007-02-28 23:59:59', 730)
-                      type: string
-                      expr: date_sub('2009-01-01', 1)
-                      type: string
-                      expr: date_sub('2009-01-01', 365)
-                      type: string
-                      expr: date_sub('2008-02-28', 2)
-                      type: string
-                      expr: date_sub('2009-02-28', 2)
-                      type: string
-                      expr: date_sub('2007-02-28', 365)
-                      type: string
-                      expr: date_sub('2007-02-28 01:12:34', 730)
-                      type: string
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              Filter Operator
+                predicate:
+                    expr: (UDFToDouble(key) = UDFToDouble(86))
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: datediff('2008-12-31', '2009-01-01')
+                        type: int
+                        expr: datediff('2008-03-01', '2008-02-28')
+                        type: int
+                        expr: datediff('2007-03-01', '2007-01-28')
+                        type: int
+                        expr: datediff('2008-03-01 23:59:59', '2008-03-02 00:00:00')
+                        type: int
+                        expr: date_add('2008-12-31', 1)
+                        type: string
+                        expr: date_add('2008-12-31', 365)
+                        type: string
+                        expr: date_add('2008-02-28', 2)
+                        type: string
+                        expr: date_add('2009-02-28', 2)
+                        type: string
+                        expr: date_add('2007-02-28', 365)
+                        type: string
+                        expr: date_add('2007-02-28 23:59:59', 730)
+                        type: string
+                        expr: date_sub('2009-01-01', 1)
+                        type: string
+                        expr: date_sub('2009-01-01', 365)
+                        type: string
+                        expr: date_sub('2008-02-28', 2)
+                        type: string
+                        expr: date_sub('2009-02-28', 2)
+                        type: string
+                        expr: date_sub('2007-02-28', 365)
+                        type: string
+                        expr: date_sub('2007-02-28 01:12:34', 730)
+                        type: string
+                  File Output Operator
+                    compressed: false
+                    GlobalTableId: 0
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
   Stage: Stage-0
     Fetch Operator
@@ -80,5 +84,5 @@
        DATE_SUB('2007-02-28', 365), DATE_SUB('2007-02-28 01:12:34', 730)
        FROM src WHERE src.key = 86
 Input: default/src
-Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/2099135137/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/126482949/10000
 -1	2	32	-1	2009-01-01	2009-12-31	2008-03-01	2009-03-02	2008-02-28	2009-02-27	2008-12-31	2008-01-02	2008-02-28	2009-02-27	2006-02-28	2005-02-28

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_10_trims.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_10_trims.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_10_trims.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_10_trims.q.out Wed Jun 24 04:05:02 2009
@@ -21,18 +21,22 @@
               predicate:
                   expr: (UDFToDouble(key) = UDFToDouble(86))
                   type: boolean
-              Select Operator
-                expressions:
-                      expr: trim(trim(trim(trim(trim(trim(trim(trim(trim(trim('  abc  '))))))))))
-                      type: string
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 1
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                      name: dest1
+              Filter Operator
+                predicate:
+                    expr: (UDFToDouble(key) = UDFToDouble(86))
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: trim(trim(trim(trim(trim(trim(trim(trim(trim(trim('  abc  '))))))))))
+                        type: string
+                  File Output Operator
+                    compressed: false
+                    GlobalTableId: 1
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                        name: dest1
 
   Stage: Stage-4
     Conditional Operator
@@ -40,10 +44,10 @@
           Move Operator
             files:
                 hdfs directory: true
-                destination: file:/data/users/njain/hive4/hive4/build/ql/tmp/561019190/10000
+                destination: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1489774447/10000
           Map Reduce
             Alias -> Map Operator Tree:
-              file:/data/users/njain/hive4/hive4/build/ql/tmp/545728748/10002 
+              file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1596742883/10002 
                   Reduce Output Operator
                     sort order: 
                     Map-reduce partition columns:

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_isnull_isnotnull.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_isnull_isnotnull.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_isnull_isnotnull.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_isnull_isnotnull.q.out Wed Jun 24 04:05:02 2009
@@ -20,21 +20,25 @@
               predicate:
                   expr: true is not null
                   type: boolean
-              Select Operator
-                expressions:
-                      expr: null is null
-                      type: boolean
-                      expr: 1 is not null
-                      type: boolean
-                      expr: 'my string' is not null
-                      type: boolean
-                Limit
-                  File Output Operator
-                    compressed: false
-                    GlobalTableId: 0
-                    table:
-                        input format: org.apache.hadoop.mapred.TextInputFormat
-                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              Filter Operator
+                predicate:
+                    expr: true is not null
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: null is null
+                        type: boolean
+                        expr: 1 is not null
+                        type: boolean
+                        expr: 'my string' is not null
+                        type: boolean
+                  Limit
+                    File Output Operator
+                      compressed: false
+                      GlobalTableId: 0
+                      table:
+                          input format: org.apache.hadoop.mapred.TextInputFormat
+                          output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
   Stage: Stage-0
     Fetch Operator
@@ -47,7 +51,7 @@
 FROM src
 WHERE true IS NOT NULL LIMIT 1
 Input: default/src
-Output: file:/home/zhoumin/hive-trunk/build/ql/tmp/1515895443/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1433540979/10000
 true	true	true
 query: EXPLAIN
 FROM src_thrift
@@ -72,21 +76,25 @@
               predicate:
                   expr: (lint is not null and not mstringstring is null)
                   type: boolean
-              Select Operator
-                expressions:
-                      expr: lint is not null
-                      type: boolean
-                      expr: lintstring is not null
-                      type: boolean
-                      expr: mstringstring is not null
-                      type: boolean
-                Limit
-                  File Output Operator
-                    compressed: false
-                    GlobalTableId: 0
-                    table:
-                        input format: org.apache.hadoop.mapred.TextInputFormat
-                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              Filter Operator
+                predicate:
+                    expr: (lint is not null and not mstringstring is null)
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: lint is not null
+                        type: boolean
+                        expr: lintstring is not null
+                        type: boolean
+                        expr: mstringstring is not null
+                        type: boolean
+                  Limit
+                    File Output Operator
+                      compressed: false
+                      GlobalTableId: 0
+                      table:
+                          input format: org.apache.hadoop.mapred.TextInputFormat
+                          output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
   Stage: Stage-0
     Fetch Operator
@@ -100,5 +108,5 @@
 WHERE  src_thrift.lint IS NOT NULL 
        AND NOT (src_thrift.mstringstring IS NULL) LIMIT 1
 Input: default/src_thrift
-Output: file:/home/zhoumin/hive-trunk/build/ql/tmp/302479509/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/63009813/10000
 true	true	true

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_like.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_like.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_like.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_like.q.out Wed Jun 24 04:05:02 2009
@@ -19,38 +19,42 @@
               predicate:
                   expr: (UDFToDouble(key) = UDFToDouble(86))
                   type: boolean
-              Select Operator
-                expressions:
-                      expr: ('_%_' like '%\_\%\_%')
-                      type: boolean
-                      expr: ('__' like '%\_\%\_%')
-                      type: boolean
-                      expr: ('%%_%_' like '%\_\%\_%')
-                      type: boolean
-                      expr: ('%_%_%' like '%\%\_\%')
-                      type: boolean
-                      expr: ('_%_' like '\%\_%')
-                      type: boolean
-                      expr: ('%__' like '__\%%')
-                      type: boolean
-                      expr: ('_%' like '\_\%\_\%%')
-                      type: boolean
-                      expr: ('_%' like '\_\%_%')
-                      type: boolean
-                      expr: ('%_' like '\%\_')
-                      type: boolean
-                      expr: ('ab' like '\%\_')
-                      type: boolean
-                      expr: ('ab' like '_a%')
-                      type: boolean
-                      expr: ('ab' like 'a')
-                      type: boolean
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              Filter Operator
+                predicate:
+                    expr: (UDFToDouble(key) = UDFToDouble(86))
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: ('_%_' like '%\_\%\_%')
+                        type: boolean
+                        expr: ('__' like '%\_\%\_%')
+                        type: boolean
+                        expr: ('%%_%_' like '%\_\%\_%')
+                        type: boolean
+                        expr: ('%_%_%' like '%\%\_\%')
+                        type: boolean
+                        expr: ('_%_' like '\%\_%')
+                        type: boolean
+                        expr: ('%__' like '__\%%')
+                        type: boolean
+                        expr: ('_%' like '\_\%\_\%%')
+                        type: boolean
+                        expr: ('_%' like '\_\%_%')
+                        type: boolean
+                        expr: ('%_' like '\%\_')
+                        type: boolean
+                        expr: ('ab' like '\%\_')
+                        type: boolean
+                        expr: ('ab' like '_a%')
+                        type: boolean
+                        expr: ('ab' like 'a')
+                        type: boolean
+                  File Output Operator
+                    compressed: false
+                    GlobalTableId: 0
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
   Stage: Stage-0
     Fetch Operator
@@ -62,5 +66,5 @@
   '%_' LIKE '\%\_', 'ab' LIKE '\%\_', 'ab' LIKE '_a%', 'ab' LIKE 'a'
 FROM src WHERE src.key = 86
 Input: default/src
-Output: file:/home/yjia/hive/build/ql/tmp/839410699/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/551411911/10000
 true	false	true	true	false	false	false	false	true	false	false	false

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_lower.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_lower.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_lower.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_lower.q.out Wed Jun 24 04:05:02 2009
@@ -16,18 +16,22 @@
               predicate:
                   expr: (UDFToDouble(key) = UDFToDouble(86))
                   type: boolean
-              Select Operator
-                expressions:
-                      expr: lower('AbC 123')
-                      type: string
-                      expr: ucase('AbC 123')
-                      type: string
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              Filter Operator
+                predicate:
+                    expr: (UDFToDouble(key) = UDFToDouble(86))
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: lower('AbC 123')
+                        type: string
+                        expr: ucase('AbC 123')
+                        type: string
+                  File Output Operator
+                    compressed: false
+                    GlobalTableId: 0
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
   Stage: Stage-0
     Fetch Operator
@@ -36,5 +40,5 @@
 
 query: SELECT lower('AbC 123'), upper('AbC 123') FROM src WHERE key = 86
 Input: default/src
-Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1926782866/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/894848839/10000
 abc 123	ABC 123

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_parse_url.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_parse_url.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_parse_url.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_parse_url.q.out Wed Jun 24 04:05:02 2009
@@ -27,36 +27,40 @@
               predicate:
                   expr: (UDFToDouble(key) = UDFToDouble(86))
                   type: boolean
-              Select Operator
-                expressions:
-                      expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'HOST')
-                      type: string
-                      expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'PATH')
-                      type: string
-                      expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY')
-                      type: string
-                      expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'REF')
-                      type: string
-                      expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY', 'k2')
-                      type: string
-                      expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY', 'k1')
-                      type: string
-                      expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY', 'k3')
-                      type: string
-                      expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'FILE')
-                      type: string
-                      expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'PROTOCOL')
-                      type: string
-                      expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'USERINFO')
-                      type: string
-                      expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'AUTHORITY')
-                      type: string
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              Filter Operator
+                predicate:
+                    expr: (UDFToDouble(key) = UDFToDouble(86))
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'HOST')
+                        type: string
+                        expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'PATH')
+                        type: string
+                        expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY')
+                        type: string
+                        expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'REF')
+                        type: string
+                        expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY', 'k2')
+                        type: string
+                        expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY', 'k1')
+                        type: string
+                        expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY', 'k3')
+                        type: string
+                        expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'FILE')
+                        type: string
+                        expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'PROTOCOL')
+                        type: string
+                        expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'USERINFO')
+                        type: string
+                        expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'AUTHORITY')
+                        type: string
+                  File Output Operator
+                    compressed: false
+                    GlobalTableId: 0
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
   Stage: Stage-0
     Fetch Operator
@@ -76,5 +80,5 @@
 parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'AUTHORITY') 
   FROM src WHERE key = 86
 Input: default/src
-Output: file:/mnt/vol/devrs008.snc1/suresh/hive_external/build/ql/tmp/384013469/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/17165920/10000
 facebook.com	/path1/p.php	k1=v1&k2=v2	Ref1	v2	v1	NULL	/path1/p.php?k1=v1&k2=v2	http	NULL	facebook.com

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_size.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_size.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_size.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_size.q.out Wed Jun 24 04:05:02 2009
@@ -22,23 +22,27 @@
               predicate:
                   expr: (lint is not null and not mstringstring is null)
                   type: boolean
-              Select Operator
-                expressions:
-                      expr: size(lint)
-                      type: int
-                      expr: size(lintstring)
-                      type: int
-                      expr: size(mstringstring)
-                      type: int
-                      expr: size(null)
-                      type: int
-                Limit
-                  File Output Operator
-                    compressed: false
-                    GlobalTableId: 0
-                    table:
-                        input format: org.apache.hadoop.mapred.TextInputFormat
-                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              Filter Operator
+                predicate:
+                    expr: (lint is not null and not mstringstring is null)
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: size(lint)
+                        type: int
+                        expr: size(lintstring)
+                        type: int
+                        expr: size(mstringstring)
+                        type: int
+                        expr: size(null)
+                        type: int
+                  Limit
+                    File Output Operator
+                      compressed: false
+                      GlobalTableId: 0
+                      table:
+                          input format: org.apache.hadoop.mapred.TextInputFormat
+                          output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
   Stage: Stage-0
     Fetch Operator
@@ -53,5 +57,5 @@
 WHERE  src_thrift.lint IS NOT NULL 
        AND NOT (src_thrift.mstringstring IS NULL) LIMIT 1
 Input: default/src_thrift
-Output: file:/home/zhoumin/hive-trunk/build/ql/tmp/639668664/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/774936511/10000
 3	1	1	-1

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/union.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/union.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/union.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/union.q.out Wed Jun 24 04:05:02 2009
@@ -24,49 +24,57 @@
               predicate:
                   expr: (UDFToDouble(key) < UDFToDouble(100))
                   type: boolean
-              Select Operator
-                expressions:
-                      expr: key
-                      type: string
-                      expr: value
-                      type: string
-                Union
-                  Select Operator
-                    expressions:
-                          expr: _col0
-                          type: string
-                          expr: _col1
-                          type: string
-                    File Output Operator
-                      compressed: false
-                      GlobalTableId: 1
-                      table:
-                          input format: org.apache.hadoop.mapred.TextInputFormat
-                          output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              Filter Operator
+                predicate:
+                    expr: (UDFToDouble(key) < UDFToDouble(100))
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: key
+                        type: string
+                        expr: value
+                        type: string
+                  Union
+                    Select Operator
+                      expressions:
+                            expr: _col0
+                            type: string
+                            expr: _col1
+                            type: string
+                      File Output Operator
+                        compressed: false
+                        GlobalTableId: 1
+                        table:
+                            input format: org.apache.hadoop.mapred.TextInputFormat
+                            output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
         null-subquery2:unioninput-subquery2:src 
             Filter Operator
               predicate:
                   expr: (UDFToDouble(key) > UDFToDouble(100))
                   type: boolean
-              Select Operator
-                expressions:
-                      expr: key
-                      type: string
-                      expr: value
-                      type: string
-                Union
-                  Select Operator
-                    expressions:
-                          expr: _col0
-                          type: string
-                          expr: _col1
-                          type: string
-                    File Output Operator
-                      compressed: false
-                      GlobalTableId: 1
-                      table:
-                          input format: org.apache.hadoop.mapred.TextInputFormat
-                          output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              Filter Operator
+                predicate:
+                    expr: (UDFToDouble(key) > UDFToDouble(100))
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: key
+                        type: string
+                        expr: value
+                        type: string
+                  Union
+                    Select Operator
+                      expressions:
+                            expr: _col0
+                            type: string
+                            expr: _col1
+                            type: string
+                      File Output Operator
+                        compressed: false
+                        GlobalTableId: 1
+                        table:
+                            input format: org.apache.hadoop.mapred.TextInputFormat
+                            output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
   Stage: Stage-4
     Conditional Operator
@@ -74,10 +82,10 @@
           Move Operator
             files:
                 hdfs directory: true
-                destination: file:/data/users/njain/hive4/hive4/build/ql/tmp/1845430872/10000
+                destination: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/888465041/10000
           Map Reduce
             Alias -> Map Operator Tree:
-              file:/data/users/njain/hive4/hive4/build/ql/tmp/1580671244/10001 
+              file:/data/users/pchakka/workspace/oshive/build/ql/tmp/976024110/10001 
                   Reduce Output Operator
                     sort order: 
                     Map-reduce partition columns:

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/union20.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/union20.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/union20.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/union20.q.out Wed Jun 24 04:05:02 2009
@@ -64,7 +64,7 @@
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1399413755/10002 
+        file:/data/users/pchakka/workspace/oshive/build/ql/tmp/579163613/10002 
           Union
             Reduce Output Operator
               key expressions:
@@ -80,7 +80,7 @@
                     type: string
                     expr: _col1
                     type: string
-        file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1399413755/10003 
+        file:/data/users/pchakka/workspace/oshive/build/ql/tmp/579163613/10003 
           Union
             Reduce Output Operator
               key expressions:
@@ -143,19 +143,23 @@
               predicate:
                   expr: (UDFToDouble(key) < UDFToDouble(10))
                   type: boolean
-              Select Operator
-                expressions:
-                      expr: key
-                      type: string
-                      expr: value
-                      type: string
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-                  table:
-                      input format: org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      name: binary_table
+              Filter Operator
+                predicate:
+                    expr: (UDFToDouble(key) < UDFToDouble(10))
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: key
+                        type: string
+                        expr: value
+                        type: string
+                  File Output Operator
+                    compressed: false
+                    GlobalTableId: 0
+                    table:
+                        input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                        name: binary_table
 
   Stage: Stage-4
     Map Reduce
@@ -165,24 +169,28 @@
               predicate:
                   expr: (UDFToDouble(key) < UDFToDouble(10))
                   type: boolean
-              Select Operator
-                expressions:
-                      expr: key
-                      type: string
-                      expr: value
-                      type: string
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-                  table:
-                      input format: org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      name: binary_table
+              Filter Operator
+                predicate:
+                    expr: (UDFToDouble(key) < UDFToDouble(10))
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: key
+                        type: string
+                        expr: value
+                        type: string
+                  File Output Operator
+                    compressed: false
+                    GlobalTableId: 0
+                    table:
+                        input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                        name: binary_table
 
   Stage: Stage-5
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1399413755/10004 
+        file:/data/users/pchakka/workspace/oshive/build/ql/tmp/579163613/10004 
           Union
             File Output Operator
               compressed: false
@@ -191,7 +199,7 @@
                   input format: org.apache.hadoop.mapred.SequenceFileInputFormat
                   output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
                   name: binary_table
-        file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1399413755/10006 
+        file:/data/users/pchakka/workspace/oshive/build/ql/tmp/579163613/10006 
           Union
             File Output Operator
               compressed: false
@@ -250,7 +258,7 @@
       select s4.key as key, s4.value as value from src s4 where s4.key < 10) unionsrc2
 ON (unionsrc1.key = unionsrc2.key)
 Input: default/src
-Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/572515511/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1483573814/10000
 0	val_0	0	val_0
 0	val_0	0	val_0
 0	val_0	0	val_0