You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by zs...@apache.org on 2009/06/24 06:05:05 UTC
svn commit: r787898 [5/19] - in /hadoop/hive/trunk: ./ conf/
ql/src/java/org/apache/hadoop/hive/ql/ppd/
ql/src/test/results/clientpositive/ ql/src/test/results/compiler/plan/
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out Wed Jun 24 04:05:02 2009
@@ -19,49 +19,53 @@
s
Filter Operator
predicate:
- expr: (((hash(key) & 2147483647) % 4) = 0)
+ expr: ((((hash(key) & 2147483647) % 4) = 0) and (UDFToDouble(key) > UDFToDouble(100)))
type: boolean
Filter Operator
predicate:
- expr: (UDFToDouble(key) > UDFToDouble(100))
+ expr: (((hash(key) & 2147483647) % 4) = 0)
type: boolean
- Select Operator
- expressions:
- expr: key
- type: string
- expr: value
- type: string
+ Filter Operator
+ predicate:
+ expr: (UDFToDouble(key) > UDFToDouble(100))
+ type: boolean
Select Operator
expressions:
- expr: UDFToInteger(_col0)
- type: int
- expr: _col1
+ expr: key
+ type: string
+ expr: value
type: string
- File Output Operator
- compressed: false
- GlobalTableId: 1
- directory: file:/data/users/njain/hive4/hive4/build/ql/tmp/862363094/10002
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- properties:
- name dest1
- columns.types int:string
- serialization.ddl struct dest1 { i32 key, string value}
- serialization.format 1
- columns key,value
- bucket_count -1
- serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/dest1
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: dest1
+ Select Operator
+ expressions:
+ expr: UDFToInteger(_col0)
+ type: int
+ expr: _col1
+ type: string
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ directory: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1062372910/10002
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ name dest1
+ columns.types int:string
+ serialization.ddl struct dest1 { i32 key, string value}
+ serialization.format 1
+ columns key,value
+ bucket_count -1
+ serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/dest1
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: dest1
Needs Tagging: false
Path -> Alias:
- file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcbucket/kv1.txt
+ file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcbucket/kv1.txt
Path -> Partition:
- file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcbucket/kv1.txt
+ file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcbucket/kv1.txt
Partition
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -77,7 +81,7 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcbucket
+ location file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcbucket
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcbucket
@@ -87,11 +91,11 @@
Move Operator
files:
hdfs directory: true
- source: file:/data/users/njain/hive4/hive4/build/ql/tmp/862363094/10002
- destination: file:/data/users/njain/hive4/hive4/build/ql/tmp/1120714236/10000
+ source: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1062372910/10002
+ destination: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/921828791/10000
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/njain/hive4/hive4/build/ql/tmp/862363094/10002
+ file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1062372910/10002
Reduce Output Operator
sort order:
Map-reduce partition columns:
@@ -105,9 +109,9 @@
type: string
Needs Tagging: false
Path -> Alias:
- file:/data/users/njain/hive4/hive4/build/ql/tmp/862363094/10002
+ file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1062372910/10002
Path -> Partition:
- file:/data/users/njain/hive4/hive4/build/ql/tmp/862363094/10002
+ file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1062372910/10002
Partition
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -122,7 +126,7 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/dest1
+ location file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/dest1
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest1
Reduce Operator Tree:
@@ -130,7 +134,7 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/njain/hive4/hive4/build/ql/tmp/1120714236/10000
+ directory: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/921828791/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -144,7 +148,7 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/dest1
+ location file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/dest1
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest1
@@ -152,7 +156,7 @@
Move Operator
tables:
replace: true
- source: file:/data/users/njain/hive4/hive4/build/ql/tmp/1120714236/10000
+ source: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/921828791/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -166,10 +170,10 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/dest1
+ location file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/dest1
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest1
- tmp directory: file:/data/users/njain/hive4/hive4/build/ql/tmp/1120714236/10001
+ tmp directory: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/921828791/10001
query: INSERT OVERWRITE TABLE dest1 SELECT s.*
@@ -179,7 +183,7 @@
Output: default/dest1
query: SELECT dest1.* FROM dest1
Input: default/dest1
-Output: file:/data/users/njain/hive4/hive4/build/ql/tmp/1094962031/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1564717662/10000
165 val_165
484 val_484
150 val_150
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/sample8.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/sample8.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/sample8.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/sample8.q.out Wed Jun 24 04:05:02 2009
@@ -23,43 +23,51 @@
predicate:
expr: (((hash(key) & 2147483647) % 10) = 0)
type: boolean
- Reduce Output Operator
- sort order:
- tag: 1
- value expressions:
- expr: key
- type: string
- expr: value
- type: string
- expr: ds
- type: string
- expr: hr
- type: string
+ Filter Operator
+ predicate:
+ expr: (((hash(key) & 2147483647) % 10) = 0)
+ type: boolean
+ Reduce Output Operator
+ sort order:
+ tag: 1
+ value expressions:
+ expr: key
+ type: string
+ expr: value
+ type: string
+ expr: ds
+ type: string
+ expr: hr
+ type: string
s
Filter Operator
predicate:
- expr: (((hash(key) & 2147483647) % 1) = 0)
+ expr: (((((((hash(key) & 2147483647) % 1) = 0) and (ds = '2008-04-08')) and (hr = '11')) and (ds = '2008-04-08')) and (hr = '11'))
type: boolean
- Reduce Output Operator
- sort order:
- tag: 0
- value expressions:
- expr: key
- type: string
- expr: value
- type: string
- expr: ds
- type: string
- expr: hr
- type: string
+ Filter Operator
+ predicate:
+ expr: (((hash(key) & 2147483647) % 1) = 0)
+ type: boolean
+ Reduce Output Operator
+ sort order:
+ tag: 0
+ value expressions:
+ expr: key
+ type: string
+ expr: value
+ type: string
+ expr: ds
+ type: string
+ expr: hr
+ type: string
Needs Tagging: true
Path -> Alias:
- file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
- file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
- file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11
- file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12
+ file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
+ file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
+ file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11
+ file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12
Path -> Partition:
- file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
+ file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
Partition
partition values:
ds 2008-04-08
@@ -78,10 +86,10 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart
+ location file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
- file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
+ file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
Partition
partition values:
ds 2008-04-08
@@ -100,10 +108,10 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart
+ location file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
- file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11
+ file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11
Partition
partition values:
ds 2008-04-09
@@ -122,10 +130,10 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart
+ location file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
- file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12
+ file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12
Partition
partition values:
ds 2008-04-09
@@ -144,7 +152,7 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/Users/char/Documents/workspace/Hive-460/build/ql/test/data/warehouse/srcpart
+ location file:/data/users/pchakka/workspace/oshive/build/ql/test/data/warehouse/srcpart
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
Reduce Operator Tree:
@@ -171,7 +179,7 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1858927587/10002
+ directory: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/709064843/10002
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -184,7 +192,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1858927587/10002
+ file:/data/users/pchakka/workspace/oshive/build/ql/tmp/709064843/10002
Reduce Output Operator
key expressions:
expr: _col0
@@ -209,9 +217,9 @@
type: string
Needs Tagging: false
Path -> Alias:
- file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1858927587/10002
+ file:/data/users/pchakka/workspace/oshive/build/ql/tmp/709064843/10002
Path -> Partition:
- file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1858927587/10002
+ file:/data/users/pchakka/workspace/oshive/build/ql/tmp/709064843/10002
Partition
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -226,7 +234,7 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1858927587/10001
+ directory: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/709064843/10001
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -249,7 +257,7 @@
Input: default/srcpart/ds=2008-04-08/hr=12
Input: default/srcpart/ds=2008-04-09/hr=11
Input: default/srcpart/ds=2008-04-09/hr=12
-Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/587483226/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/362051028/10000
0 val_0
0 val_0
0 val_0
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/subq.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/subq.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/subq.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/subq.q.out Wed Jun 24 04:05:02 2009
@@ -20,24 +20,28 @@
predicate:
expr: (UDFToDouble(key) < UDFToDouble(100))
type: boolean
- Select Operator
- expressions:
- expr: key
- type: string
- expr: value
- type: string
+ Filter Operator
+ predicate:
+ expr: (UDFToDouble(key) < UDFToDouble(100))
+ type: boolean
Select Operator
expressions:
- expr: _col0
+ expr: key
type: string
- expr: _col1
+ expr: value
type: string
- File Output Operator
- compressed: false
- GlobalTableId: 1
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ Select Operator
+ expressions:
+ expr: _col0
+ type: string
+ expr: _col1
+ type: string
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Stage: Stage-4
Conditional Operator
@@ -45,10 +49,10 @@
Move Operator
files:
hdfs directory: true
- destination: file:/data/users/njain/hive4/hive4/build/ql/tmp/260945618/10000
+ destination: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1386424262/10000
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/njain/hive4/hive4/build/ql/tmp/1316342742/10001
+ file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1220050565/10001
Reduce Output Operator
sort order:
Map-reduce partition columns:
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out Wed Jun 24 04:05:02 2009
@@ -14,29 +14,33 @@
Map Reduce
Alias -> Map Operator Tree:
a:b
- Select Operator
- expressions:
- expr: key
- type: string
- Group By Operator
- aggregations:
- expr: count(1)
- keys:
+ Filter Operator
+ predicate:
+ expr: (UDFToDouble(key) >= UDFToDouble(90))
+ type: boolean
+ Select Operator
+ expressions:
expr: key
type: string
- mode: hash
- Reduce Output Operator
- key expressions:
- expr: _col0
- type: string
- sort order: +
- Map-reduce partition columns:
- expr: _col0
+ Group By Operator
+ aggregations:
+ expr: count(1)
+ keys:
+ expr: key
type: string
- tag: -1
- value expressions:
- expr: _col1
- type: bigint
+ mode: hash
+ Reduce Output Operator
+ key expressions:
+ expr: _col0
+ type: string
+ sort order: +
+ Map-reduce partition columns:
+ expr: _col0
+ type: string
+ tag: -1
+ value expressions:
+ expr: _col1
+ type: bigint
Reduce Operator Tree:
Group By Operator
aggregations:
@@ -77,7 +81,7 @@
FROM (SELECT b.key as k, count(1) as c FROM src b GROUP BY b.key) a
WHERE a.k >= 90
Input: default/src
-Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/2044684588/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/259421857/10000
100 2
103 2
104 2
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf1.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf1.q.out Wed Jun 24 04:05:02 2009
@@ -28,56 +28,60 @@
predicate:
expr: (UDFToDouble(key) = UDFToDouble(86))
type: boolean
- Select Operator
- expressions:
- expr: ('a' like '%a%')
- type: boolean
- expr: ('b' like '%a%')
- type: boolean
- expr: ('ab' like '%a%')
- type: boolean
- expr: ('ab' like '%a_')
- type: boolean
- expr: ('%_' like '\%\_')
- type: boolean
- expr: ('ab' like '\%\_')
- type: boolean
- expr: ('ab' like '_a%')
- type: boolean
- expr: ('ab' like 'a')
- type: boolean
- expr: ('' regexp '.*')
- type: boolean
- expr: ('a' regexp '[ab]')
- type: boolean
- expr: ('' regexp '[ab]')
- type: boolean
- expr: ('hadoop' regexp '[a-z]*')
- type: boolean
- expr: ('hadoop' regexp 'o*')
- type: boolean
- expr: regexp_replace('abc', 'b', 'c')
- type: string
- expr: regexp_replace('abc', 'z', 'a')
- type: string
- expr: regexp_replace('abbbb', 'bb', 'b')
- type: string
- expr: regexp_replace('hadoop', '(.)[a-z]*', '$1ive')
- type: string
- expr: regexp_replace('hadoopAAA', 'A.*', '')
- type: string
- expr: regexp_replace('abc', '', 'A')
- type: string
- expr: ('abc' regexp '')
- type: boolean
- File Output Operator
- compressed: false
- GlobalTableId: 1
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: dest1
+ Filter Operator
+ predicate:
+ expr: (UDFToDouble(key) = UDFToDouble(86))
+ type: boolean
+ Select Operator
+ expressions:
+ expr: ('a' like '%a%')
+ type: boolean
+ expr: ('b' like '%a%')
+ type: boolean
+ expr: ('ab' like '%a%')
+ type: boolean
+ expr: ('ab' like '%a_')
+ type: boolean
+ expr: ('%_' like '\%\_')
+ type: boolean
+ expr: ('ab' like '\%\_')
+ type: boolean
+ expr: ('ab' like '_a%')
+ type: boolean
+ expr: ('ab' like 'a')
+ type: boolean
+ expr: ('' regexp '.*')
+ type: boolean
+ expr: ('a' regexp '[ab]')
+ type: boolean
+ expr: ('' regexp '[ab]')
+ type: boolean
+ expr: ('hadoop' regexp '[a-z]*')
+ type: boolean
+ expr: ('hadoop' regexp 'o*')
+ type: boolean
+ expr: regexp_replace('abc', 'b', 'c')
+ type: string
+ expr: regexp_replace('abc', 'z', 'a')
+ type: string
+ expr: regexp_replace('abbbb', 'bb', 'b')
+ type: string
+ expr: regexp_replace('hadoop', '(.)[a-z]*', '$1ive')
+ type: string
+ expr: regexp_replace('hadoopAAA', 'A.*', '')
+ type: string
+ expr: regexp_replace('abc', '', 'A')
+ type: string
+ expr: ('abc' regexp '')
+ type: boolean
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: dest1
Stage: Stage-4
Conditional Operator
@@ -85,10 +89,10 @@
Move Operator
files:
hdfs directory: true
- destination: file:/data/users/njain/hive4/hive4/build/ql/tmp/1253611032/10000
+ destination: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/734075252/10000
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/njain/hive4/hive4/build/ql/tmp/1620724991/10002
+ file:/data/users/pchakka/workspace/oshive/build/ql/tmp/816992741/10002
Reduce Output Operator
sort order:
Map-reduce partition columns:
@@ -169,5 +173,5 @@
Output: default/dest1
query: SELECT dest1.* FROM dest1
Input: default/dest1
-Output: file:/data/users/njain/hive4/hive4/build/ql/tmp/1990021744/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/674207509/10000
true false true true true false false false true true false true false acc abc abb hive hadoop AaAbAcA false
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf9.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf9.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf9.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf9.q.out Wed Jun 24 04:05:02 2009
@@ -24,46 +24,50 @@
predicate:
expr: (UDFToDouble(key) = UDFToDouble(86))
type: boolean
- Select Operator
- expressions:
- expr: datediff('2008-12-31', '2009-01-01')
- type: int
- expr: datediff('2008-03-01', '2008-02-28')
- type: int
- expr: datediff('2007-03-01', '2007-01-28')
- type: int
- expr: datediff('2008-03-01 23:59:59', '2008-03-02 00:00:00')
- type: int
- expr: date_add('2008-12-31', 1)
- type: string
- expr: date_add('2008-12-31', 365)
- type: string
- expr: date_add('2008-02-28', 2)
- type: string
- expr: date_add('2009-02-28', 2)
- type: string
- expr: date_add('2007-02-28', 365)
- type: string
- expr: date_add('2007-02-28 23:59:59', 730)
- type: string
- expr: date_sub('2009-01-01', 1)
- type: string
- expr: date_sub('2009-01-01', 365)
- type: string
- expr: date_sub('2008-02-28', 2)
- type: string
- expr: date_sub('2009-02-28', 2)
- type: string
- expr: date_sub('2007-02-28', 365)
- type: string
- expr: date_sub('2007-02-28 01:12:34', 730)
- type: string
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ Filter Operator
+ predicate:
+ expr: (UDFToDouble(key) = UDFToDouble(86))
+ type: boolean
+ Select Operator
+ expressions:
+ expr: datediff('2008-12-31', '2009-01-01')
+ type: int
+ expr: datediff('2008-03-01', '2008-02-28')
+ type: int
+ expr: datediff('2007-03-01', '2007-01-28')
+ type: int
+ expr: datediff('2008-03-01 23:59:59', '2008-03-02 00:00:00')
+ type: int
+ expr: date_add('2008-12-31', 1)
+ type: string
+ expr: date_add('2008-12-31', 365)
+ type: string
+ expr: date_add('2008-02-28', 2)
+ type: string
+ expr: date_add('2009-02-28', 2)
+ type: string
+ expr: date_add('2007-02-28', 365)
+ type: string
+ expr: date_add('2007-02-28 23:59:59', 730)
+ type: string
+ expr: date_sub('2009-01-01', 1)
+ type: string
+ expr: date_sub('2009-01-01', 365)
+ type: string
+ expr: date_sub('2008-02-28', 2)
+ type: string
+ expr: date_sub('2009-02-28', 2)
+ type: string
+ expr: date_sub('2007-02-28', 365)
+ type: string
+ expr: date_sub('2007-02-28 01:12:34', 730)
+ type: string
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Stage: Stage-0
Fetch Operator
@@ -80,5 +84,5 @@
DATE_SUB('2007-02-28', 365), DATE_SUB('2007-02-28 01:12:34', 730)
FROM src WHERE src.key = 86
Input: default/src
-Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/2099135137/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/126482949/10000
-1 2 32 -1 2009-01-01 2009-12-31 2008-03-01 2009-03-02 2008-02-28 2009-02-27 2008-12-31 2008-01-02 2008-02-28 2009-02-27 2006-02-28 2005-02-28
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_10_trims.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_10_trims.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_10_trims.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_10_trims.q.out Wed Jun 24 04:05:02 2009
@@ -21,18 +21,22 @@
predicate:
expr: (UDFToDouble(key) = UDFToDouble(86))
type: boolean
- Select Operator
- expressions:
- expr: trim(trim(trim(trim(trim(trim(trim(trim(trim(trim(' abc '))))))))))
- type: string
- File Output Operator
- compressed: false
- GlobalTableId: 1
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: dest1
+ Filter Operator
+ predicate:
+ expr: (UDFToDouble(key) = UDFToDouble(86))
+ type: boolean
+ Select Operator
+ expressions:
+ expr: trim(trim(trim(trim(trim(trim(trim(trim(trim(trim(' abc '))))))))))
+ type: string
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: dest1
Stage: Stage-4
Conditional Operator
@@ -40,10 +44,10 @@
Move Operator
files:
hdfs directory: true
- destination: file:/data/users/njain/hive4/hive4/build/ql/tmp/561019190/10000
+ destination: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1489774447/10000
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/njain/hive4/hive4/build/ql/tmp/545728748/10002
+ file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1596742883/10002
Reduce Output Operator
sort order:
Map-reduce partition columns:
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_isnull_isnotnull.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_isnull_isnotnull.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_isnull_isnotnull.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_isnull_isnotnull.q.out Wed Jun 24 04:05:02 2009
@@ -20,21 +20,25 @@
predicate:
expr: true is not null
type: boolean
- Select Operator
- expressions:
- expr: null is null
- type: boolean
- expr: 1 is not null
- type: boolean
- expr: 'my string' is not null
- type: boolean
- Limit
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ Filter Operator
+ predicate:
+ expr: true is not null
+ type: boolean
+ Select Operator
+ expressions:
+ expr: null is null
+ type: boolean
+ expr: 1 is not null
+ type: boolean
+ expr: 'my string' is not null
+ type: boolean
+ Limit
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Stage: Stage-0
Fetch Operator
@@ -47,7 +51,7 @@
FROM src
WHERE true IS NOT NULL LIMIT 1
Input: default/src
-Output: file:/home/zhoumin/hive-trunk/build/ql/tmp/1515895443/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1433540979/10000
true true true
query: EXPLAIN
FROM src_thrift
@@ -72,21 +76,25 @@
predicate:
expr: (lint is not null and not mstringstring is null)
type: boolean
- Select Operator
- expressions:
- expr: lint is not null
- type: boolean
- expr: lintstring is not null
- type: boolean
- expr: mstringstring is not null
- type: boolean
- Limit
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ Filter Operator
+ predicate:
+ expr: (lint is not null and not mstringstring is null)
+ type: boolean
+ Select Operator
+ expressions:
+ expr: lint is not null
+ type: boolean
+ expr: lintstring is not null
+ type: boolean
+ expr: mstringstring is not null
+ type: boolean
+ Limit
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Stage: Stage-0
Fetch Operator
@@ -100,5 +108,5 @@
WHERE src_thrift.lint IS NOT NULL
AND NOT (src_thrift.mstringstring IS NULL) LIMIT 1
Input: default/src_thrift
-Output: file:/home/zhoumin/hive-trunk/build/ql/tmp/302479509/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/63009813/10000
true true true
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_like.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_like.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_like.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_like.q.out Wed Jun 24 04:05:02 2009
@@ -19,38 +19,42 @@
predicate:
expr: (UDFToDouble(key) = UDFToDouble(86))
type: boolean
- Select Operator
- expressions:
- expr: ('_%_' like '%\_\%\_%')
- type: boolean
- expr: ('__' like '%\_\%\_%')
- type: boolean
- expr: ('%%_%_' like '%\_\%\_%')
- type: boolean
- expr: ('%_%_%' like '%\%\_\%')
- type: boolean
- expr: ('_%_' like '\%\_%')
- type: boolean
- expr: ('%__' like '__\%%')
- type: boolean
- expr: ('_%' like '\_\%\_\%%')
- type: boolean
- expr: ('_%' like '\_\%_%')
- type: boolean
- expr: ('%_' like '\%\_')
- type: boolean
- expr: ('ab' like '\%\_')
- type: boolean
- expr: ('ab' like '_a%')
- type: boolean
- expr: ('ab' like 'a')
- type: boolean
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ Filter Operator
+ predicate:
+ expr: (UDFToDouble(key) = UDFToDouble(86))
+ type: boolean
+ Select Operator
+ expressions:
+ expr: ('_%_' like '%\_\%\_%')
+ type: boolean
+ expr: ('__' like '%\_\%\_%')
+ type: boolean
+ expr: ('%%_%_' like '%\_\%\_%')
+ type: boolean
+ expr: ('%_%_%' like '%\%\_\%')
+ type: boolean
+ expr: ('_%_' like '\%\_%')
+ type: boolean
+ expr: ('%__' like '__\%%')
+ type: boolean
+ expr: ('_%' like '\_\%\_\%%')
+ type: boolean
+ expr: ('_%' like '\_\%_%')
+ type: boolean
+ expr: ('%_' like '\%\_')
+ type: boolean
+ expr: ('ab' like '\%\_')
+ type: boolean
+ expr: ('ab' like '_a%')
+ type: boolean
+ expr: ('ab' like 'a')
+ type: boolean
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Stage: Stage-0
Fetch Operator
@@ -62,5 +66,5 @@
'%_' LIKE '\%\_', 'ab' LIKE '\%\_', 'ab' LIKE '_a%', 'ab' LIKE 'a'
FROM src WHERE src.key = 86
Input: default/src
-Output: file:/home/yjia/hive/build/ql/tmp/839410699/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/551411911/10000
true false true true false false false false true false false false
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_lower.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_lower.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_lower.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_lower.q.out Wed Jun 24 04:05:02 2009
@@ -16,18 +16,22 @@
predicate:
expr: (UDFToDouble(key) = UDFToDouble(86))
type: boolean
- Select Operator
- expressions:
- expr: lower('AbC 123')
- type: string
- expr: ucase('AbC 123')
- type: string
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ Filter Operator
+ predicate:
+ expr: (UDFToDouble(key) = UDFToDouble(86))
+ type: boolean
+ Select Operator
+ expressions:
+ expr: lower('AbC 123')
+ type: string
+ expr: ucase('AbC 123')
+ type: string
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Stage: Stage-0
Fetch Operator
@@ -36,5 +40,5 @@
query: SELECT lower('AbC 123'), upper('AbC 123') FROM src WHERE key = 86
Input: default/src
-Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1926782866/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/894848839/10000
abc 123 ABC 123
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_parse_url.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_parse_url.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_parse_url.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_parse_url.q.out Wed Jun 24 04:05:02 2009
@@ -27,36 +27,40 @@
predicate:
expr: (UDFToDouble(key) = UDFToDouble(86))
type: boolean
- Select Operator
- expressions:
- expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'HOST')
- type: string
- expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'PATH')
- type: string
- expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY')
- type: string
- expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'REF')
- type: string
- expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY', 'k2')
- type: string
- expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY', 'k1')
- type: string
- expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY', 'k3')
- type: string
- expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'FILE')
- type: string
- expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'PROTOCOL')
- type: string
- expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'USERINFO')
- type: string
- expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'AUTHORITY')
- type: string
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ Filter Operator
+ predicate:
+ expr: (UDFToDouble(key) = UDFToDouble(86))
+ type: boolean
+ Select Operator
+ expressions:
+ expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'HOST')
+ type: string
+ expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'PATH')
+ type: string
+ expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY')
+ type: string
+ expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'REF')
+ type: string
+ expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY', 'k2')
+ type: string
+ expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY', 'k1')
+ type: string
+ expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY', 'k3')
+ type: string
+ expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'FILE')
+ type: string
+ expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'PROTOCOL')
+ type: string
+ expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'USERINFO')
+ type: string
+ expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'AUTHORITY')
+ type: string
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Stage: Stage-0
Fetch Operator
@@ -76,5 +80,5 @@
parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'AUTHORITY')
FROM src WHERE key = 86
Input: default/src
-Output: file:/mnt/vol/devrs008.snc1/suresh/hive_external/build/ql/tmp/384013469/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/17165920/10000
facebook.com /path1/p.php k1=v1&k2=v2 Ref1 v2 v1 NULL /path1/p.php?k1=v1&k2=v2 http NULL facebook.com
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_size.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_size.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_size.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_size.q.out Wed Jun 24 04:05:02 2009
@@ -22,23 +22,27 @@
predicate:
expr: (lint is not null and not mstringstring is null)
type: boolean
- Select Operator
- expressions:
- expr: size(lint)
- type: int
- expr: size(lintstring)
- type: int
- expr: size(mstringstring)
- type: int
- expr: size(null)
- type: int
- Limit
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ Filter Operator
+ predicate:
+ expr: (lint is not null and not mstringstring is null)
+ type: boolean
+ Select Operator
+ expressions:
+ expr: size(lint)
+ type: int
+ expr: size(lintstring)
+ type: int
+ expr: size(mstringstring)
+ type: int
+ expr: size(null)
+ type: int
+ Limit
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Stage: Stage-0
Fetch Operator
@@ -53,5 +57,5 @@
WHERE src_thrift.lint IS NOT NULL
AND NOT (src_thrift.mstringstring IS NULL) LIMIT 1
Input: default/src_thrift
-Output: file:/home/zhoumin/hive-trunk/build/ql/tmp/639668664/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/774936511/10000
3 1 1 -1
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/union.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/union.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/union.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/union.q.out Wed Jun 24 04:05:02 2009
@@ -24,49 +24,57 @@
predicate:
expr: (UDFToDouble(key) < UDFToDouble(100))
type: boolean
- Select Operator
- expressions:
- expr: key
- type: string
- expr: value
- type: string
- Union
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: _col1
- type: string
- File Output Operator
- compressed: false
- GlobalTableId: 1
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ Filter Operator
+ predicate:
+ expr: (UDFToDouble(key) < UDFToDouble(100))
+ type: boolean
+ Select Operator
+ expressions:
+ expr: key
+ type: string
+ expr: value
+ type: string
+ Union
+ Select Operator
+ expressions:
+ expr: _col0
+ type: string
+ expr: _col1
+ type: string
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
null-subquery2:unioninput-subquery2:src
Filter Operator
predicate:
expr: (UDFToDouble(key) > UDFToDouble(100))
type: boolean
- Select Operator
- expressions:
- expr: key
- type: string
- expr: value
- type: string
- Union
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: _col1
- type: string
- File Output Operator
- compressed: false
- GlobalTableId: 1
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ Filter Operator
+ predicate:
+ expr: (UDFToDouble(key) > UDFToDouble(100))
+ type: boolean
+ Select Operator
+ expressions:
+ expr: key
+ type: string
+ expr: value
+ type: string
+ Union
+ Select Operator
+ expressions:
+ expr: _col0
+ type: string
+ expr: _col1
+ type: string
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Stage: Stage-4
Conditional Operator
@@ -74,10 +82,10 @@
Move Operator
files:
hdfs directory: true
- destination: file:/data/users/njain/hive4/hive4/build/ql/tmp/1845430872/10000
+ destination: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/888465041/10000
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/njain/hive4/hive4/build/ql/tmp/1580671244/10001
+ file:/data/users/pchakka/workspace/oshive/build/ql/tmp/976024110/10001
Reduce Output Operator
sort order:
Map-reduce partition columns:
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/union20.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/union20.q.out?rev=787898&r1=787897&r2=787898&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/union20.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/union20.q.out Wed Jun 24 04:05:02 2009
@@ -64,7 +64,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1399413755/10002
+ file:/data/users/pchakka/workspace/oshive/build/ql/tmp/579163613/10002
Union
Reduce Output Operator
key expressions:
@@ -80,7 +80,7 @@
type: string
expr: _col1
type: string
- file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1399413755/10003
+ file:/data/users/pchakka/workspace/oshive/build/ql/tmp/579163613/10003
Union
Reduce Output Operator
key expressions:
@@ -143,19 +143,23 @@
predicate:
expr: (UDFToDouble(key) < UDFToDouble(10))
type: boolean
- Select Operator
- expressions:
- expr: key
- type: string
- expr: value
- type: string
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.SequenceFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
- name: binary_table
+ Filter Operator
+ predicate:
+ expr: (UDFToDouble(key) < UDFToDouble(10))
+ type: boolean
+ Select Operator
+ expressions:
+ expr: key
+ type: string
+ expr: value
+ type: string
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ name: binary_table
Stage: Stage-4
Map Reduce
@@ -165,24 +169,28 @@
predicate:
expr: (UDFToDouble(key) < UDFToDouble(10))
type: boolean
- Select Operator
- expressions:
- expr: key
- type: string
- expr: value
- type: string
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.SequenceFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
- name: binary_table
+ Filter Operator
+ predicate:
+ expr: (UDFToDouble(key) < UDFToDouble(10))
+ type: boolean
+ Select Operator
+ expressions:
+ expr: key
+ type: string
+ expr: value
+ type: string
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ name: binary_table
Stage: Stage-5
Map Reduce
Alias -> Map Operator Tree:
- file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1399413755/10004
+ file:/data/users/pchakka/workspace/oshive/build/ql/tmp/579163613/10004
Union
File Output Operator
compressed: false
@@ -191,7 +199,7 @@
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
name: binary_table
- file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/1399413755/10006
+ file:/data/users/pchakka/workspace/oshive/build/ql/tmp/579163613/10006
Union
File Output Operator
compressed: false
@@ -250,7 +258,7 @@
select s4.key as key, s4.value as value from src s4 where s4.key < 10) unionsrc2
ON (unionsrc1.key = unionsrc2.key)
Input: default/src
-Output: file:/Users/char/Documents/workspace/Hive-460/build/ql/tmp/572515511/10000
+Output: file:/data/users/pchakka/workspace/oshive/build/ql/tmp/1483573814/10000
0 val_0 0 val_0
0 val_0 0 val_0
0 val_0 0 val_0