You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by nz...@apache.org on 2010/02/14 03:31:44 UTC
svn commit: r909965 [6/15] - in /hadoop/hive/trunk: ./
contrib/src/test/results/clientnegative/
contrib/src/test/results/clientpositive/
ql/src/java/org/apache/hadoop/hive/ql/exec/
ql/src/test/results/clientpositive/
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/join_map_ppr.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/join_map_ppr.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/join_map_ppr.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/join_map_ppr.q.out Sun Feb 14 02:31:40 2010
@@ -84,37 +84,37 @@
File Output Operator
compressed: false
GlobalTableId: 1
- directory: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1886349005/10002
+ directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10002
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name dest_j1
+ bucket_count -1
+ columns key,value,val2
columns.types string:string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+ name dest_j1
serialization.ddl struct dest_j1 { string key, string value, string val2}
serialization.format 1
- columns key,value,val2
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
- transient_lastDdlTime 1263595165
+ transient_lastDdlTime 1266042780
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest_j1
Local Work:
Map Reduce Local Work
Alias -> Map Local Tables:
- y
+ x
Fetch Operator
limit: -1
- x
+ y
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- y
+ x
TableScan
- alias: y
+ alias: x
Common Join Operator
condition map:
Inner Join 0 to 1
@@ -160,27 +160,27 @@
File Output Operator
compressed: false
GlobalTableId: 1
- directory: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1886349005/10002
+ directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10002
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name dest_j1
+ bucket_count -1
+ columns key,value,val2
columns.types string:string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+ name dest_j1
serialization.ddl struct dest_j1 { string key, string value, string val2}
serialization.format 1
- columns key,value,val2
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
- transient_lastDdlTime 1263595165
+ transient_lastDdlTime 1266042780
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest_j1
- x
+ y
TableScan
- alias: x
+ alias: y
Common Join Operator
condition map:
Inner Join 0 to 1
@@ -226,29 +226,29 @@
File Output Operator
compressed: false
GlobalTableId: 1
- directory: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1886349005/10002
+ directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10002
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name dest_j1
+ bucket_count -1
+ columns key,value,val2
columns.types string:string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+ name dest_j1
serialization.ddl struct dest_j1 { string key, string value, string val2}
serialization.format 1
- columns key,value,val2
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
- transient_lastDdlTime 1263595165
+ transient_lastDdlTime 1266042780
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest_j1
Needs Tagging: false
Path -> Alias:
- file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [z]
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [z]
Path -> Partition:
- file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
Partition
base file name: hr=11
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -257,35 +257,35 @@
ds 2008-04-08
hr 11
properties:
- name srcpart
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart
+ name srcpart
+ partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
- columns key,value
- partition_columns ds/hr
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/srcpart
- transient_lastDdlTime 1263595164
+ transient_lastDdlTime 1266042778
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name srcpart
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart
+ name srcpart
+ partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
- columns key,value
- partition_columns ds/hr
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/srcpart
- transient_lastDdlTime 1263595164
+ transient_lastDdlTime 1266042778
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
name: srcpart
@@ -297,37 +297,37 @@
Move Operator
files:
hdfs directory: true
- source: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1886349005/10002
- destination: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/141026855/10000
+ source: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10002
+ destination: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10000
Stage: Stage-0
Move Operator
tables:
replace: true
- source: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/141026855/10000
+ source: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name dest_j1
+ bucket_count -1
+ columns key,value,val2
columns.types string:string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+ name dest_j1
serialization.ddl struct dest_j1 { string key, string value, string val2}
serialization.format 1
- columns key,value,val2
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
- transient_lastDdlTime 1263595165
+ transient_lastDdlTime 1266042780
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest_j1
- tmp directory: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/141026855/10001
+ tmp directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10001
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1886349005/10002
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10002
Reduce Output Operator
sort order:
Map-reduce partition columns:
@@ -343,41 +343,41 @@
type: string
Needs Tagging: false
Path -> Alias:
- file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1886349005/10002 [file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1886349005/10002]
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10002 [file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10002]
Path -> Partition:
- file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1886349005/10002
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10002
Partition
base file name: 10002
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name dest_j1
+ bucket_count -1
+ columns key,value,val2
columns.types string:string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+ name dest_j1
serialization.ddl struct dest_j1 { string key, string value, string val2}
serialization.format 1
- columns key,value,val2
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
- transient_lastDdlTime 1263595165
+ transient_lastDdlTime 1266042780
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name dest_j1
+ bucket_count -1
+ columns key,value,val2
columns.types string:string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+ name dest_j1
serialization.ddl struct dest_j1 { string key, string value, string val2}
serialization.format 1
- columns key,value,val2
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
- transient_lastDdlTime 1263595165
+ transient_lastDdlTime 1266042780
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest_j1
name: dest_j1
@@ -386,22 +386,22 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/141026855/10000
+ directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name dest_j1
+ bucket_count -1
+ columns key,value,val2
columns.types string:string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+ name dest_j1
serialization.ddl struct dest_j1 { string key, string value, string val2}
serialization.format 1
- columns key,value,val2
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- transient_lastDdlTime 1263595165
+ transient_lastDdlTime 1266042780
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest_j1
@@ -429,11 +429,11 @@
PREHOOK: query: select * from dest_j1 x order by x.key
PREHOOK: type: QUERY
PREHOOK: Input: default@dest_j1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/886762969/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-06_516_3338852843532728387/10000
POSTHOOK: query: select * from dest_j1 x order by x.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest_j1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/886762969/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-06_516_3338852843532728387/10000
128 val_128 val_128
128 val_128 val_128
128 val_128 val_128
@@ -648,37 +648,37 @@
File Output Operator
compressed: false
GlobalTableId: 1
- directory: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1775434766/10002
+ directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10002
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name dest_j1
+ bucket_count -1
+ columns key,value,val2
columns.types string:string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+ name dest_j1
serialization.ddl struct dest_j1 { string key, string value, string val2}
serialization.format 1
- columns key,value,val2
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
- transient_lastDdlTime 1263595165
+ transient_lastDdlTime 1266042780
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest_j1
Local Work:
Map Reduce Local Work
Alias -> Map Local Tables:
- y
+ x
Fetch Operator
limit: -1
- x
+ y
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- y
+ x
TableScan
- alias: y
+ alias: x
Common Join Operator
condition map:
Inner Join 0 to 1
@@ -724,27 +724,27 @@
File Output Operator
compressed: false
GlobalTableId: 1
- directory: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1775434766/10002
+ directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10002
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name dest_j1
+ bucket_count -1
+ columns key,value,val2
columns.types string:string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+ name dest_j1
serialization.ddl struct dest_j1 { string key, string value, string val2}
serialization.format 1
- columns key,value,val2
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
- transient_lastDdlTime 1263595165
+ transient_lastDdlTime 1266042780
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest_j1
- x
+ y
TableScan
- alias: x
+ alias: y
Common Join Operator
condition map:
Inner Join 0 to 1
@@ -790,29 +790,29 @@
File Output Operator
compressed: false
GlobalTableId: 1
- directory: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1775434766/10002
+ directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10002
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name dest_j1
+ bucket_count -1
+ columns key,value,val2
columns.types string:string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+ name dest_j1
serialization.ddl struct dest_j1 { string key, string value, string val2}
serialization.format 1
- columns key,value,val2
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
- transient_lastDdlTime 1263595165
+ transient_lastDdlTime 1266042780
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest_j1
Needs Tagging: false
Path -> Alias:
- file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [z]
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [z]
Path -> Partition:
- file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
Partition
base file name: hr=11
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -821,35 +821,35 @@
ds 2008-04-08
hr 11
properties:
- name srcpart
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart
+ name srcpart
+ partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
- columns key,value
- partition_columns ds/hr
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/srcpart
- transient_lastDdlTime 1263595164
+ transient_lastDdlTime 1266042778
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name srcpart
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart
+ name srcpart
+ partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
- columns key,value
- partition_columns ds/hr
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/srcpart
- transient_lastDdlTime 1263595164
+ transient_lastDdlTime 1266042778
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
name: srcpart
@@ -861,37 +861,37 @@
Move Operator
files:
hdfs directory: true
- source: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1775434766/10002
- destination: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1505671688/10000
+ source: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10002
+ destination: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10000
Stage: Stage-0
Move Operator
tables:
replace: true
- source: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1505671688/10000
+ source: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name dest_j1
+ bucket_count -1
+ columns key,value,val2
columns.types string:string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+ name dest_j1
serialization.ddl struct dest_j1 { string key, string value, string val2}
serialization.format 1
- columns key,value,val2
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
- transient_lastDdlTime 1263595165
+ transient_lastDdlTime 1266042780
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest_j1
- tmp directory: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1505671688/10001
+ tmp directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10001
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1775434766/10002
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10002
Reduce Output Operator
sort order:
Map-reduce partition columns:
@@ -907,41 +907,41 @@
type: string
Needs Tagging: false
Path -> Alias:
- file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1775434766/10002 [file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1775434766/10002]
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10002 [file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10002]
Path -> Partition:
- file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1775434766/10002
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10002
Partition
base file name: 10002
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name dest_j1
+ bucket_count -1
+ columns key,value,val2
columns.types string:string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+ name dest_j1
serialization.ddl struct dest_j1 { string key, string value, string val2}
serialization.format 1
- columns key,value,val2
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
- transient_lastDdlTime 1263595165
+ transient_lastDdlTime 1266042780
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name dest_j1
+ bucket_count -1
+ columns key,value,val2
columns.types string:string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+ name dest_j1
serialization.ddl struct dest_j1 { string key, string value, string val2}
serialization.format 1
- columns key,value,val2
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
- transient_lastDdlTime 1263595165
+ transient_lastDdlTime 1266042780
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest_j1
name: dest_j1
@@ -950,22 +950,22 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1505671688/10000
+ directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name dest_j1
+ bucket_count -1
+ columns key,value,val2
columns.types string:string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+ name dest_j1
serialization.ddl struct dest_j1 { string key, string value, string val2}
serialization.format 1
- columns key,value,val2
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- transient_lastDdlTime 1263595165
+ transient_lastDdlTime 1266042780
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest_j1
@@ -993,11 +993,11 @@
PREHOOK: query: select * from dest_j1 x order by x.key
PREHOOK: type: QUERY
PREHOOK: Input: default@dest_j1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/877753191/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-26_697_820182467019293566/10000
POSTHOOK: query: select * from dest_j1 x order by x.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest_j1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/877753191/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-26_697_820182467019293566/10000
128 val_128 val_128
128 val_128 val_128
128 val_128 val_128
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/join_rc.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/join_rc.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/join_rc.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/join_rc.q.out Sun Feb 14 02:31:40 2010
@@ -51,9 +51,9 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- join_rc2
+ join_rc1
TableScan
- alias: join_rc2
+ alias: join_rc1
Reduce Output Operator
key expressions:
expr: key
@@ -62,13 +62,13 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 1
+ tag: 0
value expressions:
- expr: value
+ expr: key
type: string
- join_rc1
+ join_rc2
TableScan
- alias: join_rc1
+ alias: join_rc2
Reduce Output Operator
key expressions:
expr: key
@@ -77,9 +77,9 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 0
+ tag: 1
value expressions:
- expr: key
+ expr: value
type: string
Reduce Operator Tree:
Join Operator
@@ -114,13 +114,13 @@
PREHOOK: type: QUERY
PREHOOK: Input: default@join_rc2
PREHOOK: Input: default@join_rc1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1138840055/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-27-47_640_1441550174396831657/10000
POSTHOOK: query: select join_rc1.key, join_rc2.value
FROM join_rc1 JOIN join_rc2 ON join_rc1.key = join_rc2.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@join_rc2
POSTHOOK: Input: default@join_rc1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1138840055/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-27-47_640_1441550174396831657/10000
0 val_0
0 val_0
0 val_0
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder.q.out Sun Feb 14 02:31:40 2010
@@ -57,38 +57,38 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- c
+ a
TableScan
- alias: c
+ alias: a
Reduce Output Operator
key expressions:
- expr: (key + 1)
+ expr: UDFToDouble(key)
type: double
sort order: +
Map-reduce partition columns:
- expr: (key + 1)
+ expr: UDFToDouble(key)
type: double
- tag: 1
+ tag: 0
value expressions:
expr: key
type: string
- a
+ expr: val
+ type: string
+ c
TableScan
- alias: a
+ alias: c
Reduce Output Operator
key expressions:
- expr: UDFToDouble(key)
+ expr: (key + 1)
type: double
sort order: +
Map-reduce partition columns:
- expr: UDFToDouble(key)
+ expr: (key + 1)
type: double
- tag: 0
+ tag: 1
value expressions:
expr: key
type: string
- expr: val
- type: string
Reduce Operator Tree:
Join Operator
condition map:
@@ -136,38 +136,38 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- c
+ a
TableScan
- alias: c
+ alias: a
Reduce Output Operator
key expressions:
- expr: (key + 1)
+ expr: UDFToDouble(key)
type: double
sort order: +
Map-reduce partition columns:
- expr: (key + 1)
+ expr: UDFToDouble(key)
type: double
- tag: 0
+ tag: 1
value expressions:
expr: key
type: string
- a
+ expr: val
+ type: string
+ c
TableScan
- alias: a
+ alias: c
Reduce Output Operator
key expressions:
- expr: UDFToDouble(key)
+ expr: (key + 1)
type: double
sort order: +
Map-reduce partition columns:
- expr: UDFToDouble(key)
+ expr: (key + 1)
type: double
- tag: 1
+ tag: 0
value expressions:
expr: key
type: string
- expr: val
- type: string
Reduce Operator Tree:
Join Operator
condition map:
@@ -203,13 +203,13 @@
PREHOOK: type: QUERY
PREHOOK: Input: default@src
PREHOOK: Input: default@t1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1966956520/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-27-56_522_2427479678690495417/10000
POSTHOOK: query: FROM T1 a JOIN src c ON c.key+1=a.key
SELECT a.key, a.val, c.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
POSTHOOK: Input: default@t1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1966956520/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-27-56_522_2427479678690495417/10000
1 11 0
1 11 0
1 11 0
@@ -219,13 +219,13 @@
PREHOOK: type: QUERY
PREHOOK: Input: default@src
PREHOOK: Input: default@t1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/2092123428/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-28-01_499_4645186819885871483/10000
POSTHOOK: query: FROM T1 a JOIN src c ON c.key+1=a.key
SELECT /*+ STREAMTABLE(a) */ a.key, a.val, c.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
POSTHOOK: Input: default@t1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/2092123428/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-28-01_499_4645186819885871483/10000
1 11 0
1 11 0
1 11 0
@@ -252,9 +252,9 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- b
+ a
TableScan
- alias: b
+ alias: a
Reduce Output Operator
key expressions:
expr: key
@@ -263,13 +263,15 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 1
+ tag: 0
value expressions:
expr: key
type: string
- a
+ expr: val
+ type: string
+ b
TableScan
- alias: a
+ alias: b
Reduce Output Operator
key expressions:
expr: key
@@ -278,12 +280,10 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 0
+ tag: 1
value expressions:
expr: key
type: string
- expr: val
- type: string
Reduce Operator Tree:
Join Operator
condition map:
@@ -389,9 +389,9 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- b
+ a
TableScan
- alias: b
+ alias: a
Reduce Output Operator
key expressions:
expr: key
@@ -400,13 +400,15 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 0
+ tag: 1
value expressions:
expr: key
type: string
- a
+ expr: val
+ type: string
+ b
TableScan
- alias: a
+ alias: b
Reduce Output Operator
key expressions:
expr: key
@@ -415,12 +417,10 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 1
+ tag: 0
value expressions:
expr: key
type: string
- expr: val
- type: string
Reduce Operator Tree:
Join Operator
condition map:
@@ -512,7 +512,7 @@
PREHOOK: Input: default@t2
PREHOOK: Input: default@t3
PREHOOK: Input: default@t1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1966669209/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-28-07_922_2294786151582488765/10000
POSTHOOK: query: FROM T1 a
LEFT OUTER JOIN T2 b ON (b.key=a.key)
RIGHT OUTER JOIN T3 c ON (c.val = a.val)
@@ -521,7 +521,7 @@
POSTHOOK: Input: default@t2
POSTHOOK: Input: default@t3
POSTHOOK: Input: default@t1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1966669209/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-28-07_922_2294786151582488765/10000
2 2 12 12
NULL NULL NULL 14
NULL NULL NULL 16
@@ -534,7 +534,7 @@
PREHOOK: Input: default@t2
PREHOOK: Input: default@t3
PREHOOK: Input: default@t1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/2040128444/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-28-20_301_8638454003102682278/10000
POSTHOOK: query: FROM T1 a
LEFT OUTER JOIN T2 b ON (b.key=a.key)
RIGHT OUTER JOIN T3 c ON (c.val = a.val)
@@ -543,7 +543,7 @@
POSTHOOK: Input: default@t2
POSTHOOK: Input: default@t3
POSTHOOK: Input: default@t1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/2040128444/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-28-20_301_8638454003102682278/10000
2 2 12 12
NULL NULL NULL 14
NULL NULL NULL 16
@@ -571,9 +571,9 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- b
+ a
TableScan
- alias: b
+ alias: a
Reduce Output Operator
key expressions:
expr: key
@@ -586,13 +586,13 @@
type: string
expr: val
type: string
- tag: 1
+ tag: 0
value expressions:
expr: key
type: string
- c
+ b
TableScan
- alias: c
+ alias: b
Reduce Output Operator
key expressions:
expr: key
@@ -605,13 +605,13 @@
type: string
expr: val
type: string
- tag: 2
+ tag: 1
value expressions:
expr: key
type: string
- a
+ c
TableScan
- alias: a
+ alias: c
Reduce Output Operator
key expressions:
expr: key
@@ -624,7 +624,7 @@
type: string
expr: val
type: string
- tag: 0
+ tag: 2
value expressions:
expr: key
type: string
@@ -683,9 +683,9 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- b
+ a
TableScan
- alias: b
+ alias: a
Reduce Output Operator
key expressions:
expr: key
@@ -698,13 +698,13 @@
type: string
expr: val
type: string
- tag: 2
+ tag: 0
value expressions:
expr: key
type: string
- c
+ b
TableScan
- alias: c
+ alias: b
Reduce Output Operator
key expressions:
expr: key
@@ -717,13 +717,13 @@
type: string
expr: val
type: string
- tag: 1
+ tag: 2
value expressions:
expr: key
type: string
- a
+ c
TableScan
- alias: a
+ alias: c
Reduce Output Operator
key expressions:
expr: key
@@ -736,7 +736,7 @@
type: string
expr: val
type: string
- tag: 0
+ tag: 1
value expressions:
expr: key
type: string
@@ -781,7 +781,7 @@
PREHOOK: Input: default@t2
PREHOOK: Input: default@t3
PREHOOK: Input: default@t1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/685649783/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-28-31_833_1632343405751503469/10000
POSTHOOK: query: FROM UNIQUEJOIN
PRESERVE T1 a (a.key, a.val),
PRESERVE T2 b (b.key, b.val),
@@ -791,7 +791,7 @@
POSTHOOK: Input: default@t2
POSTHOOK: Input: default@t3
POSTHOOK: Input: default@t1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/685649783/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-28-31_833_1632343405751503469/10000
1 NULL NULL
2 NULL 2
NULL 2 NULL
@@ -812,7 +812,7 @@
PREHOOK: Input: default@t2
PREHOOK: Input: default@t3
PREHOOK: Input: default@t1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1834964794/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-28-38_626_3188343531104185080/10000
POSTHOOK: query: FROM UNIQUEJOIN
PRESERVE T1 a (a.key, a.val),
PRESERVE T2 b (b.key, b.val),
@@ -822,7 +822,7 @@
POSTHOOK: Input: default@t2
POSTHOOK: Input: default@t3
POSTHOOK: Input: default@t1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1834964794/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-28-38_626_3188343531104185080/10000
1 NULL NULL
2 NULL 2
NULL 2 NULL
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder2.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder2.q.out Sun Feb 14 02:31:40 2010
@@ -77,9 +77,9 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- d
+ a
TableScan
- alias: d
+ alias: a
Reduce Output Operator
key expressions:
expr: key
@@ -88,7 +88,7 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 0
+ tag: 3
value expressions:
expr: key
type: string
@@ -128,9 +128,9 @@
type: string
expr: val
type: string
- a
+ d
TableScan
- alias: a
+ alias: d
Reduce Output Operator
key expressions:
expr: key
@@ -139,7 +139,7 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 3
+ tag: 0
value expressions:
expr: key
type: string
@@ -198,7 +198,7 @@
PREHOOK: Input: default@t2
PREHOOK: Input: default@t3
PREHOOK: Input: default@t1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/1499291437/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-28-02_101_1898487455882506837/10000
POSTHOOK: query: SELECT /*+ STREAMTABLE(a) */ *
FROM T1 a JOIN T2 b ON a.key = b.key
JOIN T3 c ON b.key = c.key
@@ -208,7 +208,7 @@
POSTHOOK: Input: default@t2
POSTHOOK: Input: default@t3
POSTHOOK: Input: default@t1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/1499291437/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-28-02_101_1898487455882506837/10000
2 12 2 22 2 12 2 12
PREHOOK: query: EXPLAIN
SELECT /*+ STREAMTABLE(a) */ *
@@ -235,9 +235,9 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- b
+ a
TableScan
- alias: b
+ alias: a
Reduce Output Operator
key expressions:
expr: key
@@ -246,15 +246,15 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 0
+ tag: 1
value expressions:
expr: key
type: string
expr: val
type: string
- a
+ b
TableScan
- alias: a
+ alias: b
Reduce Output Operator
key expressions:
expr: key
@@ -263,7 +263,7 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 1
+ tag: 0
value expressions:
expr: key
type: string
@@ -343,23 +343,6 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- d
- TableScan
- alias: d
- Reduce Output Operator
- key expressions:
- expr: (key + 1)
- type: double
- sort order: +
- Map-reduce partition columns:
- expr: (key + 1)
- type: double
- tag: 0
- value expressions:
- expr: key
- type: string
- expr: val
- type: string
$INTNAME
Reduce Output Operator
key expressions:
@@ -383,6 +366,23 @@
type: string
expr: _col3
type: string
+ d
+ TableScan
+ alias: d
+ Reduce Output Operator
+ key expressions:
+ expr: (key + 1)
+ type: double
+ sort order: +
+ Map-reduce partition columns:
+ expr: (key + 1)
+ type: double
+ tag: 0
+ value expressions:
+ expr: key
+ type: string
+ expr: val
+ type: string
Reduce Operator Tree:
Join Operator
condition map:
@@ -431,7 +431,7 @@
PREHOOK: Input: default@t2
PREHOOK: Input: default@t3
PREHOOK: Input: default@t1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/347443381/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-28-09_022_5817489663400235377/10000
POSTHOOK: query: SELECT /*+ STREAMTABLE(a) */ *
FROM T1 a JOIN T2 b ON a.key = b.key
JOIN T3 c ON a.val = c.val
@@ -441,7 +441,7 @@
POSTHOOK: Input: default@t2
POSTHOOK: Input: default@t3
POSTHOOK: Input: default@t1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/347443381/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-28-09_022_5817489663400235377/10000
2 22 2 12 2 12 2 12
PREHOOK: query: DROP TABLE T1
PREHOOK: type: DROPTABLE
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder3.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder3.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder3.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder3.q.out Sun Feb 14 02:31:40 2010
@@ -77,9 +77,9 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- d
+ a
TableScan
- alias: d
+ alias: a
Reduce Output Operator
key expressions:
expr: key
@@ -88,7 +88,7 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 0
+ tag: 3
value expressions:
expr: key
type: string
@@ -128,9 +128,9 @@
type: string
expr: val
type: string
- a
+ d
TableScan
- alias: a
+ alias: d
Reduce Output Operator
key expressions:
expr: key
@@ -139,7 +139,7 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 3
+ tag: 0
value expressions:
expr: key
type: string
@@ -198,7 +198,7 @@
PREHOOK: Input: default@t2
PREHOOK: Input: default@t3
PREHOOK: Input: default@t1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/922482225/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-34_907_1935969450376975821/10000
POSTHOOK: query: SELECT /*+ STREAMTABLE(a,c) */ *
FROM T1 a JOIN T2 b ON a.key = b.key
JOIN T3 c ON b.key = c.key
@@ -208,7 +208,7 @@
POSTHOOK: Input: default@t2
POSTHOOK: Input: default@t3
POSTHOOK: Input: default@t1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/922482225/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-34_907_1935969450376975821/10000
2 12 2 22 2 12 2 12
PREHOOK: query: EXPLAIN
SELECT /*+ STREAMTABLE(a,c) */ *
@@ -235,9 +235,9 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- b
+ a
TableScan
- alias: b
+ alias: a
Reduce Output Operator
key expressions:
expr: key
@@ -246,15 +246,15 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 0
+ tag: 1
value expressions:
expr: key
type: string
expr: val
type: string
- a
+ b
TableScan
- alias: a
+ alias: b
Reduce Output Operator
key expressions:
expr: key
@@ -263,7 +263,7 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 1
+ tag: 0
value expressions:
expr: key
type: string
@@ -343,23 +343,6 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- d
- TableScan
- alias: d
- Reduce Output Operator
- key expressions:
- expr: (key + 1)
- type: double
- sort order: +
- Map-reduce partition columns:
- expr: (key + 1)
- type: double
- tag: 0
- value expressions:
- expr: key
- type: string
- expr: val
- type: string
$INTNAME
Reduce Output Operator
key expressions:
@@ -383,6 +366,23 @@
type: string
expr: _col3
type: string
+ d
+ TableScan
+ alias: d
+ Reduce Output Operator
+ key expressions:
+ expr: (key + 1)
+ type: double
+ sort order: +
+ Map-reduce partition columns:
+ expr: (key + 1)
+ type: double
+ tag: 0
+ value expressions:
+ expr: key
+ type: string
+ expr: val
+ type: string
Reduce Operator Tree:
Join Operator
condition map:
@@ -431,7 +431,7 @@
PREHOOK: Input: default@t2
PREHOOK: Input: default@t3
PREHOOK: Input: default@t1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/831264775/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-41_586_5996486802122601404/10000
POSTHOOK: query: SELECT /*+ STREAMTABLE(a,c) */ *
FROM T1 a JOIN T2 b ON a.key = b.key
JOIN T3 c ON a.val = c.val
@@ -441,7 +441,7 @@
POSTHOOK: Input: default@t2
POSTHOOK: Input: default@t3
POSTHOOK: Input: default@t1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/831264775/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-41_586_5996486802122601404/10000
2 22 2 12 2 12 2 12
PREHOOK: query: DROP TABLE T1
PREHOOK: type: DROPTABLE
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/join_thrift.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/join_thrift.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/join_thrift.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/join_thrift.q.out Sun Feb 14 02:31:40 2010
@@ -31,9 +31,9 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- s2
+ s1
TableScan
- alias: s2
+ alias: s1
Reduce Output Operator
key expressions:
expr: aint
@@ -42,13 +42,13 @@
Map-reduce partition columns:
expr: aint
type: int
- tag: 1
+ tag: 0
value expressions:
- expr: lintstring
- type: array<struct<myint:int,mystring:string,underscore_int:int>>
- s1
+ expr: aint
+ type: int
+ s2
TableScan
- alias: s1
+ alias: s2
Reduce Output Operator
key expressions:
expr: aint
@@ -57,10 +57,10 @@
Map-reduce partition columns:
expr: aint
type: int
- tag: 0
+ tag: 1
value expressions:
- expr: aint
- type: int
+ expr: lintstring
+ type: array<struct<myint:int,mystring:string,underscore_int:int>>
Reduce Operator Tree:
Join Operator
condition map:
@@ -95,14 +95,14 @@
ON s1.aint = s2.aint
PREHOOK: type: QUERY
PREHOOK: Input: default@src_thrift
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/1362622837/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-28-28_139_5468299142078849499/10000
POSTHOOK: query: SELECT s1.aint, s2.lintstring
FROM src_thrift s1
JOIN src_thrift s2
ON s1.aint = s2.aint
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src_thrift
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/1362622837/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-28-28_139_5468299142078849499/10000
-1952710710 [{"myint":25,"mystring":"125","underscore_int":5}]
-1461153973 [{"myint":49,"mystring":"343","underscore_int":7}]
-751827638 [{"myint":4,"mystring":"8","underscore_int":2}]