You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by zs...@apache.org on 2008/12/17 11:35:46 UTC
svn commit: r727337 [3/3] - in /hadoop/hive/trunk: ./
ql/src/java/org/apache/hadoop/hive/ql/exec/
ql/src/java/org/apache/hadoop/hive/ql/metadata/
ql/src/java/org/apache/hadoop/hive/ql/parse/
ql/src/java/org/apache/hadoop/hive/ql/plan/ ql/src/test/queri...
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce6.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce6.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce6.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce6.q.out Wed Dec 17 02:35:44 2008
@@ -44,13 +44,23 @@
type: string
Reduce Operator Tree:
Extract
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
- name: dest1
+ Select Operator
+ expressions:
+ expr: UDFToInteger(0)
+ type: int
+ expr: 1
+ type: int
+ expr: 2
+ type: int
+ expr: 3
+ type: string
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+ name: dest1
Stage: Stage-0
Move Operator
@@ -59,7 +69,7 @@
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+ serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
name: dest1
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/notable_alias1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/notable_alias1.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/notable_alias1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/notable_alias1.q.out Wed Dec 17 02:35:44 2008
@@ -49,7 +49,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- /tmp/hive-jssarma/238416396/213974514.10001
+ /tmp/hive-zshao/10020404/161743246.10001
Reduce Output Operator
key expressions:
expr: 0
@@ -69,7 +69,7 @@
keys:
expr: KEY.0
type: string
- mode: unknown
+ mode: final
Select Operator
expressions:
expr: '1234'
@@ -78,13 +78,21 @@
type: string
expr: 1
type: bigint
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
- name: dest1
+ Select Operator
+ expressions:
+ expr: 0
+ type: string
+ expr: UDFToInteger(1)
+ type: int
+ expr: UDFToDouble(2)
+ type: double
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+ name: dest1
Stage: Stage-0
Move Operator
@@ -93,64 +101,64 @@
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+ serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
name: dest1
-1234 0 3
-1234 10 1
-1234 11 1
-1234 12 2
-1234 15 2
-1234 17 1
-1234 18 2
-1234 19 1
-1234 2 1
-1234 20 1
-1234 24 2
-1234 26 2
-1234 27 1
-1234 28 1
-1234 30 1
-1234 33 1
-1234 34 1
-1234 35 3
-1234 37 2
-1234 4 1
-1234 41 1
-1234 42 2
-1234 43 1
-1234 44 1
-1234 47 1
-1234 5 3
-1234 51 2
-1234 53 1
-1234 54 1
-1234 57 1
-1234 58 2
-1234 64 1
-1234 65 1
-1234 66 1
-1234 67 2
-1234 69 1
-1234 70 3
-1234 72 2
-1234 74 1
-1234 76 2
-1234 77 1
-1234 78 1
-1234 8 1
-1234 80 1
-1234 82 1
-1234 83 2
-1234 84 2
-1234 85 1
-1234 86 1
-1234 87 1
-1234 9 1
-1234 90 3
-1234 92 1
-1234 95 2
-1234 96 1
-1234 97 2
-1234 98 2
+1234 0 3.0
+1234 10 1.0
+1234 11 1.0
+1234 12 2.0
+1234 15 2.0
+1234 17 1.0
+1234 18 2.0
+1234 19 1.0
+1234 2 1.0
+1234 20 1.0
+1234 24 2.0
+1234 26 2.0
+1234 27 1.0
+1234 28 1.0
+1234 30 1.0
+1234 33 1.0
+1234 34 1.0
+1234 35 3.0
+1234 37 2.0
+1234 4 1.0
+1234 41 1.0
+1234 42 2.0
+1234 43 1.0
+1234 44 1.0
+1234 47 1.0
+1234 5 3.0
+1234 51 2.0
+1234 53 1.0
+1234 54 1.0
+1234 57 1.0
+1234 58 2.0
+1234 64 1.0
+1234 65 1.0
+1234 66 1.0
+1234 67 2.0
+1234 69 1.0
+1234 70 3.0
+1234 72 2.0
+1234 74 1.0
+1234 76 2.0
+1234 77 1.0
+1234 78 1.0
+1234 8 1.0
+1234 80 1.0
+1234 82 1.0
+1234 83 2.0
+1234 84 2.0
+1234 85 1.0
+1234 86 1.0
+1234 87 1.0
+1234 9 1.0
+1234 90 3.0
+1234 92 1.0
+1234 95 2.0
+1234 96 1.0
+1234 97 2.0
+1234 98 2.0
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/notable_alias2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/notable_alias2.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/notable_alias2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/notable_alias2.q.out Wed Dec 17 02:35:44 2008
@@ -49,7 +49,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- /tmp/hive-jssarma/506681419/141455468.10001
+ /tmp/hive-zshao/55994326/602553370.10001
Reduce Output Operator
key expressions:
expr: 0
@@ -69,7 +69,7 @@
keys:
expr: KEY.0
type: string
- mode: unknown
+ mode: final
Select Operator
expressions:
expr: '1234'
@@ -78,13 +78,21 @@
type: string
expr: 1
type: bigint
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
- name: dest1
+ Select Operator
+ expressions:
+ expr: 0
+ type: string
+ expr: UDFToInteger(1)
+ type: int
+ expr: UDFToDouble(2)
+ type: double
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+ name: dest1
Stage: Stage-0
Move Operator
@@ -93,64 +101,64 @@
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+ serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
name: dest1
-1234 0 3
-1234 10 1
-1234 11 1
-1234 12 2
-1234 15 2
-1234 17 1
-1234 18 2
-1234 19 1
-1234 2 1
-1234 20 1
-1234 24 2
-1234 26 2
-1234 27 1
-1234 28 1
-1234 30 1
-1234 33 1
-1234 34 1
-1234 35 3
-1234 37 2
-1234 4 1
-1234 41 1
-1234 42 2
-1234 43 1
-1234 44 1
-1234 47 1
-1234 5 3
-1234 51 2
-1234 53 1
-1234 54 1
-1234 57 1
-1234 58 2
-1234 64 1
-1234 65 1
-1234 66 1
-1234 67 2
-1234 69 1
-1234 70 3
-1234 72 2
-1234 74 1
-1234 76 2
-1234 77 1
-1234 78 1
-1234 8 1
-1234 80 1
-1234 82 1
-1234 83 2
-1234 84 2
-1234 85 1
-1234 86 1
-1234 87 1
-1234 9 1
-1234 90 3
-1234 92 1
-1234 95 2
-1234 96 1
-1234 97 2
-1234 98 2
+1234 0 3.0
+1234 10 1.0
+1234 11 1.0
+1234 12 2.0
+1234 15 2.0
+1234 17 1.0
+1234 18 2.0
+1234 19 1.0
+1234 2 1.0
+1234 20 1.0
+1234 24 2.0
+1234 26 2.0
+1234 27 1.0
+1234 28 1.0
+1234 30 1.0
+1234 33 1.0
+1234 34 1.0
+1234 35 3.0
+1234 37 2.0
+1234 4 1.0
+1234 41 1.0
+1234 42 2.0
+1234 43 1.0
+1234 44 1.0
+1234 47 1.0
+1234 5 3.0
+1234 51 2.0
+1234 53 1.0
+1234 54 1.0
+1234 57 1.0
+1234 58 2.0
+1234 64 1.0
+1234 65 1.0
+1234 66 1.0
+1234 67 2.0
+1234 69 1.0
+1234 70 3.0
+1234 72 2.0
+1234 74 1.0
+1234 76 2.0
+1234 77 1.0
+1234 78 1.0
+1234 8 1.0
+1234 80 1.0
+1234 82 1.0
+1234 83 2.0
+1234 84 2.0
+1234 85 1.0
+1234 86 1.0
+1234 87 1.0
+1234 9 1.0
+1234 90 3.0
+1234 92 1.0
+1234 95 2.0
+1234 96 1.0
+1234 97 2.0
+1234 98 2.0
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/quote1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/quote1.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/quote1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/quote1.q.out Wed Dec 17 02:35:44 2008
@@ -20,13 +20,19 @@
type: string
expr: value
type: string
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
- name: dest1
+ Select Operator
+ expressions:
+ expr: UDFToInteger(0)
+ type: int
+ expr: 1
+ type: string
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+ name: dest1
Stage: Stage-0
Move Operator
@@ -37,7 +43,7 @@
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+ serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
name: dest1
@@ -60,7 +66,7 @@
Select Operator
expressions:
expr: location
- type: string
+ type: int
expr: type
type: string
expr: table
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/sample1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/sample1.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/sample1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/sample1.q.out Wed Dec 17 02:35:44 2008
@@ -28,29 +28,39 @@
type: string
expr: hr
type: string
- File Output Operator
- compressed: false
- directory: /tmp/hive-jssarma/37742099/114802476.10000.insclause-0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- properties:
- name dest1
- serialization.ddl struct dest1 { i32 key, string value, string dt, string hr}
- serialization.format 1
- columns key,value,dt,hr
- bucket_count -1
- serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
- serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
- name: dest1
+ Select Operator
+ expressions:
+ expr: UDFToInteger(0)
+ type: int
+ expr: 1
+ type: string
+ expr: 2
+ type: string
+ expr: 3
+ type: string
+ File Output Operator
+ compressed: false
+ directory: /tmp/hive-zshao/860547727/34027035.10000.insclause-0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+ properties:
+ name dest1
+ serialization.ddl struct dest1 { i32 key, string value, string dt, string hr}
+ serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
+ columns key,value,dt,hr
+ bucket_count -1
+ serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+ serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+ name: dest1
Needs Tagging: false
Path -> Alias:
- file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
+ file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
Path -> Partition:
- file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
+ file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
Partition
partition values:
ds 2008-04-08
@@ -68,7 +78,7 @@
serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcpart
+ location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcpart
serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
name: srcpart
@@ -76,21 +86,21 @@
Move Operator
tables:
replace: true
- source: /tmp/hive-jssarma/37742099/114802476.10000.insclause-0
+ source: /tmp/hive-zshao/860547727/34027035.10000.insclause-0
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
properties:
name dest1
serialization.ddl struct dest1 { i32 key, string value, string dt, string hr}
- serialization.format 1
+ serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
columns key,value,dt,hr
bucket_count -1
- serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+ serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
- serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+ location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+ serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
name: dest1
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/sample2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/sample2.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/sample2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/sample2.q.out Wed Dec 17 02:35:44 2008
@@ -16,29 +16,35 @@
type: string
expr: value
type: string
- File Output Operator
- compressed: false
- directory: /tmp/hive-jssarma/515988900/822291761.10000.insclause-0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- properties:
- name dest1
- serialization.ddl struct dest1 { i32 key, string value}
- serialization.format 1
- columns key,value
- bucket_count -1
- serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
- serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
- name: dest1
+ Select Operator
+ expressions:
+ expr: UDFToInteger(0)
+ type: int
+ expr: 1
+ type: string
+ File Output Operator
+ compressed: false
+ directory: /tmp/hive-zshao/427867399/126053246.10000.insclause-0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+ properties:
+ name dest1
+ serialization.ddl struct dest1 { i32 key, string value}
+ serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
+ columns key,value
+ bucket_count -1
+ serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+ serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+ name: dest1
Needs Tagging: false
Path -> Alias:
- file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket/kv1.txt
+ file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket/kv1.txt
Path -> Partition:
- file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket/kv1.txt
+ file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket/kv1.txt
Partition
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -53,7 +59,7 @@
serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket
+ location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket
serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
name: srcbucket
@@ -61,21 +67,21 @@
Move Operator
tables:
replace: true
- source: /tmp/hive-jssarma/515988900/822291761.10000.insclause-0
+ source: /tmp/hive-zshao/427867399/126053246.10000.insclause-0
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
properties:
name dest1
serialization.ddl struct dest1 { i32 key, string value}
- serialization.format 1
+ serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
columns key,value
bucket_count -1
- serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+ serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
- serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+ location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+ serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
name: dest1
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/sample4.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/sample4.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/sample4.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/sample4.q.out Wed Dec 17 02:35:44 2008
@@ -16,29 +16,35 @@
type: string
expr: value
type: string
- File Output Operator
- compressed: false
- directory: /tmp/hive-jssarma/1539308576/222195283.10000.insclause-0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- properties:
- name dest1
- serialization.ddl struct dest1 { i32 key, string value}
- serialization.format 1
- columns key,value
- bucket_count -1
- serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
- serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
- name: dest1
+ Select Operator
+ expressions:
+ expr: UDFToInteger(0)
+ type: int
+ expr: 1
+ type: string
+ File Output Operator
+ compressed: false
+ directory: /tmp/hive-zshao/134470671/579412318.10000.insclause-0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+ properties:
+ name dest1
+ serialization.ddl struct dest1 { i32 key, string value}
+ serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
+ columns key,value
+ bucket_count -1
+ serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+ serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+ name: dest1
Needs Tagging: false
Path -> Alias:
- file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket/kv1.txt
+ file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket/kv1.txt
Path -> Partition:
- file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket/kv1.txt
+ file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket/kv1.txt
Partition
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -53,7 +59,7 @@
serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket
+ location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket
serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
name: srcbucket
@@ -61,21 +67,21 @@
Move Operator
tables:
replace: true
- source: /tmp/hive-jssarma/1539308576/222195283.10000.insclause-0
+ source: /tmp/hive-zshao/134470671/579412318.10000.insclause-0
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
properties:
name dest1
serialization.ddl struct dest1 { i32 key, string value}
- serialization.format 1
+ serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
columns key,value
bucket_count -1
- serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+ serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
- serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+ location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+ serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
name: dest1
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/sample5.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/sample5.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/sample5.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/sample5.q.out Wed Dec 17 02:35:44 2008
@@ -20,29 +20,35 @@
type: string
expr: value
type: string
- File Output Operator
- compressed: false
- directory: /tmp/hive-jssarma/611887541/74229442.10000.insclause-0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- properties:
- name dest1
- serialization.ddl struct dest1 { i32 key, string value}
- serialization.format 1
- columns key,value
- bucket_count -1
- serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
- serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
- name: dest1
+ Select Operator
+ expressions:
+ expr: UDFToInteger(0)
+ type: int
+ expr: 1
+ type: string
+ File Output Operator
+ compressed: false
+ directory: /tmp/hive-zshao/211838955/159322479.10000.insclause-0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+ properties:
+ name dest1
+ serialization.ddl struct dest1 { i32 key, string value}
+ serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
+ columns key,value
+ bucket_count -1
+ serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+ serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+ name: dest1
Needs Tagging: false
Path -> Alias:
- file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket
+ file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket
Path -> Partition:
- file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket
+ file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket
Partition
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -57,7 +63,7 @@
serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket
+ location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket
serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
name: srcbucket
@@ -65,21 +71,21 @@
Move Operator
tables:
replace: true
- source: /tmp/hive-jssarma/611887541/74229442.10000.insclause-0
+ source: /tmp/hive-zshao/211838955/159322479.10000.insclause-0
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
properties:
name dest1
serialization.ddl struct dest1 { i32 key, string value}
- serialization.format 1
+ serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
columns key,value
bucket_count -1
- serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+ serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
- serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+ location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+ serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
name: dest1
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/sample6.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/sample6.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/sample6.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/sample6.q.out Wed Dec 17 02:35:44 2008
@@ -20,29 +20,35 @@
type: string
expr: value
type: string
- File Output Operator
- compressed: false
- directory: /tmp/hive-jssarma/862051292/1235999863.10000.insclause-0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- properties:
- name dest1
- serialization.ddl struct dest1 { i32 key, string value}
- serialization.format 1
- columns key,value
- bucket_count -1
- serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
- serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
- name: dest1
+ Select Operator
+ expressions:
+ expr: UDFToInteger(0)
+ type: int
+ expr: 1
+ type: string
+ File Output Operator
+ compressed: false
+ directory: /tmp/hive-zshao/25298408/261228818.10000.insclause-0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+ properties:
+ name dest1
+ serialization.ddl struct dest1 { i32 key, string value}
+ serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
+ columns key,value
+ bucket_count -1
+ serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+ serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+ name: dest1
Needs Tagging: false
Path -> Alias:
- file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket/kv1.txt
+ file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket/kv1.txt
Path -> Partition:
- file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket/kv1.txt
+ file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket/kv1.txt
Partition
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -57,7 +63,7 @@
serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket
+ location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket
serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
name: srcbucket
@@ -65,21 +71,21 @@
Move Operator
tables:
replace: true
- source: /tmp/hive-jssarma/862051292/1235999863.10000.insclause-0
+ source: /tmp/hive-zshao/25298408/261228818.10000.insclause-0
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
properties:
name dest1
serialization.ddl struct dest1 { i32 key, string value}
- serialization.format 1
+ serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
columns key,value
bucket_count -1
- serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+ serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
- serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+ location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+ serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
name: dest1
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out Wed Dec 17 02:35:44 2008
@@ -24,29 +24,35 @@
type: string
expr: value
type: string
- File Output Operator
- compressed: false
- directory: /tmp/hive-jssarma/429463772/457590891.10000.insclause-0
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- properties:
- name dest1
- serialization.ddl struct dest1 { i32 key, string value}
- serialization.format 1
- columns key,value
- bucket_count -1
- serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
- serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
- name: dest1
+ Select Operator
+ expressions:
+ expr: UDFToInteger(0)
+ type: int
+ expr: 1
+ type: string
+ File Output Operator
+ compressed: false
+ directory: /tmp/hive-zshao/444520495/211065154.10000.insclause-0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+ properties:
+ name dest1
+ serialization.ddl struct dest1 { i32 key, string value}
+ serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
+ columns key,value
+ bucket_count -1
+ serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+ serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+ name: dest1
Needs Tagging: false
Path -> Alias:
- file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket/kv1.txt
+ file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket/kv1.txt
Path -> Partition:
- file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket/kv1.txt
+ file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket/kv1.txt
Partition
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -61,7 +67,7 @@
serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket
+ location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket
serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
name: srcbucket
@@ -69,21 +75,21 @@
Move Operator
tables:
replace: true
- source: /tmp/hive-jssarma/429463772/457590891.10000.insclause-0
+ source: /tmp/hive-zshao/444520495/211065154.10000.insclause-0
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
properties:
name dest1
serialization.ddl struct dest1 { i32 key, string value}
- serialization.format 1
+ serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
columns key,value
bucket_count -1
- serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+ serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
- serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+ location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+ serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
name: dest1
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out Wed Dec 17 02:35:44 2008
@@ -45,7 +45,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- /tmp/hive-jssarma/438302216/1483130688.10002
+ /tmp/hive-zshao/368989435/823759952.10002
Reduce Output Operator
key expressions:
expr: 0
@@ -65,7 +65,7 @@
keys:
expr: KEY.0
type: string
- mode: unknown
+ mode: final
Select Operator
expressions:
expr: 0
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf3.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf3.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf3.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf3.q.out Wed Dec 17 02:35:44 2008
@@ -40,7 +40,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- /tmp/hive-jssarma/1415018212/1021817615.10001
+ /tmp/hive-zshao/546652046/1647731334.10001
Reduce Output Operator
sort order:
tag: -1
@@ -63,7 +63,7 @@
expr: count(VALUE.2)
expr: sum(VALUE.3)
expr: min(VALUE.4)
- mode: unknown
+ mode: final
Select Operator
expressions:
expr: 2
@@ -71,7 +71,7 @@
expr: 3
type: double
expr: 1
- type: string
+ type: double
expr: 4
type: double
expr: 0
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/union2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/union2.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/union2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/union2.q.out Wed Dec 17 02:35:44 2008
@@ -57,7 +57,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- /tmp/hive-jssarma/32027196/47394207.10002
+ /tmp/hive-zshao/175321401/121204523.10002
Reduce Output Operator
sort order:
tag: -1
@@ -68,7 +68,7 @@
Group By Operator
aggregations:
expr: count(VALUE.0)
- mode: unknown
+ mode: final
Select Operator
expressions:
expr: 0
Modified: hadoop/hive/trunk/ql/src/test/results/compiler/errors/invalid_index.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/compiler/errors/invalid_index.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/compiler/errors/invalid_index.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/compiler/errors/invalid_index.q.out Wed Dec 17 02:35:44 2008
@@ -1,2 +1,2 @@
Semantic Exception:
-line 2:36 [] not Valid on Non Collection Types 0
\ No newline at end of file
+line 2:36 [] not Valid on Non Collection Types 0: string
\ No newline at end of file
Modified: hadoop/hive/trunk/ql/src/test/results/compiler/plan/groupby3.q.xml
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/compiler/plan/groupby3.q.xml?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/compiler/plan/groupby3.q.xml (original)
+++ hadoop/hive/trunk/ql/src/test/results/compiler/plan/groupby3.q.xml Wed Dec 17 02:35:44 2008
@@ -20,7 +20,7 @@
<void property="aliasToWork">
<object class="java.util.HashMap">
<void method="put">
- <string>/tmp/hive-njain/332856862/73535516.10002</string>
+ <string>/tmp/hive-zshao/1268121/343796403.10002</string>
<object id="ReduceSinkOperator0" class="org.apache.hadoop.hive.ql.exec.ReduceSinkOperator">
<void property="conf">
<object class="org.apache.hadoop.hive.ql.plan.reduceSinkDesc">
@@ -176,7 +176,7 @@
<void property="conf">
<object class="org.apache.hadoop.hive.ql.plan.fileSinkDesc">
<void property="dirName">
- <string>/tmp/hive-njain/332856862/73535516.10002</string>
+ <string>/tmp/hive-zshao/1268121/343796403.10002</string>
</void>
<void property="tableInfo">
<object id="tableDesc2" class="org.apache.hadoop.hive.ql.plan.tableDesc">
@@ -853,10 +853,10 @@
<void property="pathToAliases">
<object class="java.util.LinkedHashMap">
<void method="put">
- <string>/tmp/hive-njain/332856862/73535516.10002</string>
+ <string>/tmp/hive-zshao/1268121/343796403.10002</string>
<object class="java.util.ArrayList">
<void method="add">
- <string>/tmp/hive-njain/332856862/73535516.10002</string>
+ <string>/tmp/hive-zshao/1268121/343796403.10002</string>
</void>
</object>
</void>
@@ -865,7 +865,7 @@
<void property="pathToPartitionInfo">
<object class="java.util.LinkedHashMap">
<void method="put">
- <string>/tmp/hive-njain/332856862/73535516.10002</string>
+ <string>/tmp/hive-zshao/1268121/343796403.10002</string>
<object class="org.apache.hadoop.hive.ql.plan.partitionDesc">
<void property="tableDesc">
<object idref="tableDesc2"/>
@@ -887,7 +887,7 @@
<void property="conf">
<object class="org.apache.hadoop.hive.ql.plan.fileSinkDesc">
<void property="dirName">
- <string>/tmp/hive-njain/229394155.10001.insclause-0</string>
+ <string>/tmp/hive-zshao/147248168.10001.insclause-0</string>
</void>
<void property="tableInfo">
<object class="org.apache.hadoop.hive.ql.plan.tableDesc">
@@ -943,7 +943,7 @@
<string>1</string>
</void>
<void property="type">
- <object idref="PrimitiveTypeInfo0"/>
+ <object idref="PrimitiveTypeInfo1"/>
</void>
</object>
</void>
@@ -953,7 +953,7 @@
<string>2</string>
</void>
<void property="type">
- <object idref="PrimitiveTypeInfo0"/>
+ <object idref="PrimitiveTypeInfo1"/>
</void>
</object>
</void>
@@ -1005,7 +1005,7 @@
<string>2</string>
</void>
<void property="typeInfo">
- <object idref="PrimitiveTypeInfo0"/>
+ <object idref="PrimitiveTypeInfo1"/>
</void>
</object>
</void>
@@ -1015,7 +1015,7 @@
<string>0</string>
</void>
<void property="typeInfo">
- <object idref="PrimitiveTypeInfo0"/>
+ <object idref="PrimitiveTypeInfo1"/>
</void>
</object>
</void>
@@ -1199,7 +1199,7 @@
<string>0</string>
</void>
<void property="type">
- <object idref="PrimitiveTypeInfo0"/>
+ <object idref="PrimitiveTypeInfo1"/>
</void>
</object>
</void>
@@ -1219,7 +1219,7 @@
<string>2</string>
</void>
<void property="type">
- <object idref="PrimitiveTypeInfo0"/>
+ <object idref="PrimitiveTypeInfo1"/>
</void>
</object>
</void>
@@ -1281,7 +1281,7 @@
<void property="pathToAliases">
<object class="java.util.LinkedHashMap">
<void method="put">
- <string>file:/home/njain/workspace/hadoophive/trunk/build/ql/test/data/warehouse/src</string>
+ <string>file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/src</string>
<object class="java.util.ArrayList">
<void method="add">
<string>src</string>
@@ -1293,7 +1293,7 @@
<void property="pathToPartitionInfo">
<object class="java.util.LinkedHashMap">
<void method="put">
- <string>file:/home/njain/workspace/hadoophive/trunk/build/ql/test/data/warehouse/src</string>
+ <string>file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/src</string>
<object class="org.apache.hadoop.hive.ql.plan.partitionDesc">
<void property="partSpec">
<object class="java.util.LinkedHashMap"/>
@@ -1345,7 +1345,7 @@
</void>
<void method="put">
<string>location</string>
- <string>file:/home/njain/workspace/hadoophive/trunk/build/ql/test/data/warehouse/src</string>
+ <string>file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/src</string>
</void>
</object>
</void>
Modified: hadoop/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/thrift/TCTLSeparatedProtocol.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/thrift/TCTLSeparatedProtocol.java?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/thrift/TCTLSeparatedProtocol.java (original)
+++ hadoop/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/thrift/TCTLSeparatedProtocol.java Wed Dec 17 02:35:44 2008
@@ -266,11 +266,11 @@
*/
public TCTLSeparatedProtocol(TTransport trans) {
- this(trans, defaultPrimarySeparator, defaultSecondarySeparator, defaultMapSeparator, defaultRowSeparator, false, 4096);
+ this(trans, defaultPrimarySeparator, defaultSecondarySeparator, defaultMapSeparator, defaultRowSeparator, true, 4096);
}
public TCTLSeparatedProtocol(TTransport trans, int buffer_size) {
- this(trans, defaultPrimarySeparator, defaultSecondarySeparator, defaultMapSeparator, defaultRowSeparator, false, buffer_size);
+ this(trans, defaultPrimarySeparator, defaultSecondarySeparator, defaultMapSeparator, defaultRowSeparator, true, buffer_size);
}
/**
@@ -287,8 +287,7 @@
int bufferSize) {
super(trans);
- returnNulls = returnNulls;
-
+ this.returnNulls = returnNulls;
this.primarySeparator = primarySeparator;
this.secondarySeparator = secondarySeparator;
@@ -702,31 +701,56 @@
public byte readByte() throws TException {
String val = readString();
lastPrimitiveWasNullFlag = val == null;
- return val == null || val.isEmpty() ? 0 : Byte.valueOf(val).byteValue();
+ try {
+ return val == null || val.isEmpty() ? 0 : Byte.valueOf(val).byteValue();
+ } catch (NumberFormatException e) {
+ lastPrimitiveWasNullFlag = true;
+ return 0;
+ }
}
public short readI16() throws TException {
String val = readString();
lastPrimitiveWasNullFlag = val == null;
- return val == null || val.isEmpty() ? 0 : Short.valueOf(val).shortValue();
+ try {
+ return val == null || val.isEmpty() ? 0 : Short.valueOf(val).shortValue();
+ } catch (NumberFormatException e) {
+ lastPrimitiveWasNullFlag = true;
+ return 0;
+ }
}
public int readI32() throws TException {
String val = readString();
lastPrimitiveWasNullFlag = val == null;
- return val == null || val.isEmpty() ? 0 : Integer.valueOf(val).intValue();
+ try {
+ return val == null || val.isEmpty() ? 0 : Integer.valueOf(val).intValue();
+ } catch (NumberFormatException e) {
+ lastPrimitiveWasNullFlag = true;
+ return 0;
+ }
}
public long readI64() throws TException {
String val = readString();
lastPrimitiveWasNullFlag = val == null;
- return val == null || val.isEmpty() ? 0 : Long.valueOf(val).longValue();
+ try {
+ return val == null || val.isEmpty() ? 0 : Long.valueOf(val).longValue();
+ } catch (NumberFormatException e) {
+ lastPrimitiveWasNullFlag = true;
+ return 0;
+ }
}
public double readDouble() throws TException {
String val = readString();
lastPrimitiveWasNullFlag = val == null;
- return val == null || val.isEmpty() ? 0 :Double.valueOf(val).doubleValue();
+ try {
+ return val == null || val.isEmpty() ? 0 :Double.valueOf(val).doubleValue();
+ } catch (NumberFormatException e) {
+ lastPrimitiveWasNullFlag = true;
+ return 0;
+ }
}
public String readString() throws TException {