You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by jv...@apache.org on 2010/07/30 08:40:11 UTC
svn commit: r980659 [15/34] - in /hadoop/hive/trunk: ./
common/src/java/org/apache/hadoop/hive/common/
contrib/src/test/results/clientpositive/ metastore/if/ metastore/src/gen-cpp/
metastore/src/gen-javabean/org/apache/hadoop/hive/metastore/api/ metast...
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin_negative2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin_negative2.q.out?rev=980659&r1=980658&r2=980659&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin_negative2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin_negative2.q.out Fri Jul 30 06:40:04 2010
@@ -82,7 +82,7 @@ STAGE PLANS:
keys:
0 [Column[key]]
1 [Column[key]]
- outputColumnNames: _col0, _col1, _col3
+ outputColumnNames: _col0, _col1, _col5
Position of Big Table: 0
Select Operator
expressions:
@@ -90,22 +90,22 @@ STAGE PLANS:
type: int
expr: _col1
type: string
- expr: _col3
+ expr: _col5
type: string
- outputColumnNames: _col0, _col1, _col3
+ outputColumnNames: _col0, _col1, _col5
Select Operator
expressions:
expr: _col0
type: int
expr: _col1
type: string
- expr: _col3
+ expr: _col5
type: string
outputColumnNames: _col0, _col1, _col2
File Output Operator
compressed: false
GlobalTableId: 1
- directory: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_11-46-06_937_7973298254467464329/-ext-10002
+ directory: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_11-02-14_630_2201972795156035816/-ext-10002
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -116,12 +116,12 @@ STAGE PLANS:
columns.types string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
+ location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
name bucketmapjoin_tmp_result
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1280083566
+ transient_lastDdlTime 1280426534
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucketmapjoin_tmp_result
TotalFiles: 1
@@ -146,7 +146,7 @@ STAGE PLANS:
keys:
0 [Column[key]]
1 [Column[key]]
- outputColumnNames: _col0, _col1, _col3
+ outputColumnNames: _col0, _col1, _col5
Position of Big Table: 0
Select Operator
expressions:
@@ -154,22 +154,22 @@ STAGE PLANS:
type: int
expr: _col1
type: string
- expr: _col3
+ expr: _col5
type: string
- outputColumnNames: _col0, _col1, _col3
+ outputColumnNames: _col0, _col1, _col5
Select Operator
expressions:
expr: _col0
type: int
expr: _col1
type: string
- expr: _col3
+ expr: _col5
type: string
outputColumnNames: _col0, _col1, _col2
File Output Operator
compressed: false
GlobalTableId: 1
- directory: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_11-46-06_937_7973298254467464329/-ext-10002
+ directory: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_11-02-14_630_2201972795156035816/-ext-10002
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -180,21 +180,21 @@ STAGE PLANS:
columns.types string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
+ location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
name bucketmapjoin_tmp_result
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1280083566
+ transient_lastDdlTime 1280426534
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucketmapjoin_tmp_result
TotalFiles: 1
MultiFileSpray: false
Needs Tagging: false
Path -> Alias:
- pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin [a]
+ pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/srcbucket_mapjoin [a]
Path -> Partition:
- pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin
+ pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/srcbucket_mapjoin
Partition
base file name: srcbucket_mapjoin
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -206,12 +206,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin
+ location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/srcbucket_mapjoin
name srcbucket_mapjoin
serialization.ddl struct srcbucket_mapjoin { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1280083563
+ transient_lastDdlTime 1280426530
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -223,12 +223,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin
+ location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/srcbucket_mapjoin
name srcbucket_mapjoin
serialization.ddl struct srcbucket_mapjoin { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1280083563
+ transient_lastDdlTime 1280426530
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcbucket_mapjoin
name: srcbucket_mapjoin
@@ -240,14 +240,14 @@ STAGE PLANS:
Move Operator
files:
hdfs directory: true
- source: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_11-46-06_937_7973298254467464329/-ext-10002
- destination: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_11-46-06_937_7973298254467464329/-ext-10000
+ source: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_11-02-14_630_2201972795156035816/-ext-10002
+ destination: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_11-02-14_630_2201972795156035816/-ext-10000
Stage: Stage-0
Move Operator
tables:
replace: true
- source: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_11-46-06_937_7973298254467464329/-ext-10000
+ source: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_11-02-14_630_2201972795156035816/-ext-10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -257,20 +257,20 @@ STAGE PLANS:
columns.types string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
+ location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
name bucketmapjoin_tmp_result
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1280083566
+ transient_lastDdlTime 1280426534
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucketmapjoin_tmp_result
- tmp directory: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_11-46-06_937_7973298254467464329/-ext-10001
+ tmp directory: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_11-02-14_630_2201972795156035816/-ext-10001
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_11-46-06_937_7973298254467464329/-ext-10002
+ pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_11-02-14_630_2201972795156035816/-ext-10002
Reduce Output Operator
sort order:
Map-reduce partition columns:
@@ -286,9 +286,9 @@ STAGE PLANS:
type: string
Needs Tagging: false
Path -> Alias:
- pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_11-46-06_937_7973298254467464329/-ext-10002 [pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_11-46-06_937_7973298254467464329/-ext-10002]
+ pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_11-02-14_630_2201972795156035816/-ext-10002 [pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_11-02-14_630_2201972795156035816/-ext-10002]
Path -> Partition:
- pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_11-46-06_937_7973298254467464329/-ext-10002
+ pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_11-02-14_630_2201972795156035816/-ext-10002
Partition
base file name: -ext-10002
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -299,12 +299,12 @@ STAGE PLANS:
columns.types string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
+ location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
name bucketmapjoin_tmp_result
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1280083566
+ transient_lastDdlTime 1280426534
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -315,12 +315,12 @@ STAGE PLANS:
columns.types string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
+ location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
name bucketmapjoin_tmp_result
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1280083566
+ transient_lastDdlTime 1280426534
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucketmapjoin_tmp_result
name: bucketmapjoin_tmp_result
@@ -329,7 +329,7 @@ STAGE PLANS:
File Output Operator
compressed: false
GlobalTableId: 0
- directory: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_11-46-06_937_7973298254467464329/-ext-10000
+ directory: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_11-02-14_630_2201972795156035816/-ext-10000
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -340,12 +340,12 @@ STAGE PLANS:
columns.types string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
+ location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
name bucketmapjoin_tmp_result
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1280083566
+ transient_lastDdlTime 1280426534
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucketmapjoin_tmp_result
TotalFiles: 1
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/cluster.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/cluster.q.out?rev=980659&r1=980658&r2=980659&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/cluster.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/cluster.q.out Fri Jul 30 06:40:04 2010
@@ -64,11 +64,11 @@ STAGE PLANS:
PREHOOK: query: SELECT * FROM SRC x where x.key = 10 CLUSTER BY x.key
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-30_987_946002006407811313/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-02-24_261_6437725192448955894/-mr-10000
POSTHOOK: query: SELECT * FROM SRC x where x.key = 10 CLUSTER BY x.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-30_987_946002006407811313/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-02-24_261_6437725192448955894/-mr-10000
10 val_10
PREHOOK: query: EXPLAIN
SELECT * FROM SRC x where x.key = 20 CLUSTER BY key
@@ -136,11 +136,11 @@ STAGE PLANS:
PREHOOK: query: SELECT * FROM SRC x where x.key = 20 CLUSTER BY key
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-36_798_7722926056682905425/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-02-29_842_578607839779563869/-mr-10000
POSTHOOK: query: SELECT * FROM SRC x where x.key = 20 CLUSTER BY key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-36_798_7722926056682905425/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-02-29_842_578607839779563869/-mr-10000
20 val_20
PREHOOK: query: EXPLAIN
SELECT x.* FROM SRC x where x.key = 20 CLUSTER BY key
@@ -208,11 +208,11 @@ STAGE PLANS:
PREHOOK: query: SELECT x.* FROM SRC x where x.key = 20 CLUSTER BY key
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-43_717_6012279607867421749/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-02-34_215_6197686627919208530/-mr-10000
POSTHOOK: query: SELECT x.* FROM SRC x where x.key = 20 CLUSTER BY key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-43_717_6012279607867421749/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-02-34_215_6197686627919208530/-mr-10000
20 val_20
PREHOOK: query: EXPLAIN
SELECT x.* FROM SRC x where x.key = 20 CLUSTER BY x.key
@@ -280,11 +280,11 @@ STAGE PLANS:
PREHOOK: query: SELECT x.* FROM SRC x where x.key = 20 CLUSTER BY x.key
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-49_882_6526844180884182893/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-02-38_139_3316356060320173122/-mr-10000
POSTHOOK: query: SELECT x.* FROM SRC x where x.key = 20 CLUSTER BY x.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-49_882_6526844180884182893/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-02-38_139_3316356060320173122/-mr-10000
20 val_20
PREHOOK: query: EXPLAIN
SELECT x.key, x.value as v1 FROM SRC x where x.key = 20 CLUSTER BY key
@@ -352,11 +352,11 @@ STAGE PLANS:
PREHOOK: query: SELECT x.key, x.value as v1 FROM SRC x where x.key = 20 CLUSTER BY key
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-54_276_5516587198210822350/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-02-44_253_1539937897915785116/-mr-10000
POSTHOOK: query: SELECT x.key, x.value as v1 FROM SRC x where x.key = 20 CLUSTER BY key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-54_276_5516587198210822350/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-02-44_253_1539937897915785116/-mr-10000
20 val_20
PREHOOK: query: EXPLAIN
SELECT x.key, x.value as v1 FROM SRC x where x.key = 20 CLUSTER BY x.key
@@ -424,11 +424,11 @@ STAGE PLANS:
PREHOOK: query: SELECT x.key, x.value as v1 FROM SRC x where x.key = 20 CLUSTER BY x.key
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-59_582_505878830166802854/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-02-49_068_5172852982314475051/-mr-10000
POSTHOOK: query: SELECT x.key, x.value as v1 FROM SRC x where x.key = 20 CLUSTER BY x.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-59_582_505878830166802854/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-02-49_068_5172852982314475051/-mr-10000
20 val_20
PREHOOK: query: EXPLAIN
SELECT x.key, x.value as v1 FROM SRC x where x.key = 20 CLUSTER BY v1
@@ -496,11 +496,11 @@ STAGE PLANS:
PREHOOK: query: SELECT x.key, x.value as v1 FROM SRC x where x.key = 20 CLUSTER BY v1
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-04_613_2939682790554351791/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-02-55_898_460454971604760187/-mr-10000
POSTHOOK: query: SELECT x.key, x.value as v1 FROM SRC x where x.key = 20 CLUSTER BY v1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-04_613_2939682790554351791/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-02-55_898_460454971604760187/-mr-10000
20 val_20
PREHOOK: query: EXPLAIN
SELECT y.* from (SELECT x.* FROM SRC x CLUSTER BY x.key) y where y.key = 20
@@ -571,11 +571,11 @@ STAGE PLANS:
PREHOOK: query: SELECT y.* from (SELECT x.* FROM SRC x CLUSTER BY x.key) y where y.key = 20
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-10_209_6387108226324326472/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-03-01_741_1830256943791082152/-mr-10000
POSTHOOK: query: SELECT y.* from (SELECT x.* FROM SRC x CLUSTER BY x.key) y where y.key = 20
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-10_209_6387108226324326472/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-03-01_741_1830256943791082152/-mr-10000
20 val_20
PREHOOK: query: EXPLAIN
SELECT x.key, x.value as v1, y.key FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY v1
@@ -639,7 +639,7 @@ STAGE PLANS:
0 {VALUE._col0} {VALUE._col1}
1 {VALUE._col0}
handleSkewJoin: false
- outputColumnNames: _col0, _col1, _col2
+ outputColumnNames: _col0, _col1, _col4
Filter Operator
predicate:
expr: (_col0 = 20)
@@ -650,7 +650,7 @@ STAGE PLANS:
type: string
expr: _col1
type: string
- expr: _col2
+ expr: _col4
type: string
outputColumnNames: _col0, _col1, _col2
File Output Operator
@@ -663,7 +663,7 @@ STAGE PLANS:
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-16_334_2245182201779291081/10002
+ file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-03-10_135_8465951951718830602/-mr-10002
Reduce Output Operator
key expressions:
expr: _col1
@@ -697,11 +697,11 @@ STAGE PLANS:
PREHOOK: query: SELECT x.key, x.value as v1, y.key FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY v1
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-16_594_5072146069194835371/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-03-10_286_7953538941949300288/-mr-10000
POSTHOOK: query: SELECT x.key, x.value as v1, y.key FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY v1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-16_594_5072146069194835371/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-03-10_286_7953538941949300288/-mr-10000
20 val_20 20
PREHOOK: query: EXPLAIN
SELECT x.key, x.value as v1, y.* FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY v1
@@ -767,7 +767,7 @@ STAGE PLANS:
0 {VALUE._col0} {VALUE._col1}
1 {VALUE._col0} {VALUE._col1}
handleSkewJoin: false
- outputColumnNames: _col0, _col1, _col2, _col3
+ outputColumnNames: _col0, _col1, _col4, _col5
Filter Operator
predicate:
expr: (_col0 = 20)
@@ -778,9 +778,9 @@ STAGE PLANS:
type: string
expr: _col1
type: string
- expr: _col2
+ expr: _col4
type: string
- expr: _col3
+ expr: _col5
type: string
outputColumnNames: _col0, _col1, _col2, _col3
File Output Operator
@@ -793,7 +793,7 @@ STAGE PLANS:
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-27_948_1334626995431478594/10002
+ file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-03-20_550_4090222363071736776/-mr-10002
Reduce Output Operator
key expressions:
expr: _col1
@@ -829,11 +829,11 @@ STAGE PLANS:
PREHOOK: query: SELECT x.key, x.value as v1, y.* FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY v1
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-28_074_5978892602968822143/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-03-20_636_1232432883221421266/-mr-10000
POSTHOOK: query: SELECT x.key, x.value as v1, y.* FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY v1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-28_074_5978892602968822143/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-03-20_636_1232432883221421266/-mr-10000
20 val_20 20 val_20
PREHOOK: query: EXPLAIN
SELECT x.key, x.value as v1, y.* FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY x.key
@@ -899,7 +899,7 @@ STAGE PLANS:
0 {VALUE._col0} {VALUE._col1}
1 {VALUE._col0} {VALUE._col1}
handleSkewJoin: false
- outputColumnNames: _col0, _col1, _col2, _col3
+ outputColumnNames: _col0, _col1, _col4, _col5
Filter Operator
predicate:
expr: (_col0 = 20)
@@ -910,9 +910,9 @@ STAGE PLANS:
type: string
expr: _col1
type: string
- expr: _col2
+ expr: _col4
type: string
- expr: _col3
+ expr: _col5
type: string
outputColumnNames: _col0, _col1, _col2, _col3
File Output Operator
@@ -925,7 +925,7 @@ STAGE PLANS:
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-38_372_3122728094689915036/10002
+ file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-03-29_670_6551249949062057949/-mr-10002
Reduce Output Operator
key expressions:
expr: _col0
@@ -961,11 +961,11 @@ STAGE PLANS:
PREHOOK: query: SELECT x.key, x.value as v1, y.* FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY x.key
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-38_466_4207748081161799368/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-03-29_792_7964360624661468857/-mr-10000
POSTHOOK: query: SELECT x.key, x.value as v1, y.* FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY x.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-38_466_4207748081161799368/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-03-29_792_7964360624661468857/-mr-10000
20 val_20 20 val_20
PREHOOK: query: EXPLAIN
SELECT x.key, x.value as v1, y.key as yk FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY key
@@ -1029,7 +1029,7 @@ STAGE PLANS:
0 {VALUE._col0} {VALUE._col1}
1 {VALUE._col0}
handleSkewJoin: false
- outputColumnNames: _col0, _col1, _col2
+ outputColumnNames: _col0, _col1, _col4
Filter Operator
predicate:
expr: (_col0 = 20)
@@ -1040,7 +1040,7 @@ STAGE PLANS:
type: string
expr: _col1
type: string
- expr: _col2
+ expr: _col4
type: string
outputColumnNames: _col0, _col1, _col2
File Output Operator
@@ -1053,7 +1053,7 @@ STAGE PLANS:
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-47_950_4989736117672479003/10002
+ file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-03-40_175_7869691048043021342/-mr-10002
Reduce Output Operator
key expressions:
expr: _col0
@@ -1087,11 +1087,11 @@ STAGE PLANS:
PREHOOK: query: SELECT x.key, x.value as v1, y.key as yk FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY key
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-48_073_5743203719731796760/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-03-40_269_8619996170626222765/-mr-10000
POSTHOOK: query: SELECT x.key, x.value as v1, y.key as yk FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-48_073_5743203719731796760/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-03-40_269_8619996170626222765/-mr-10000
20 val_20 20
PREHOOK: query: EXPLAIN
SELECT unioninput.*
@@ -1225,7 +1225,7 @@ FROM (
CLUSTER BY unioninput.key
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-57_574_7730562904986793585/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-03-50_185_4510032647777990341/-mr-10000
POSTHOOK: query: SELECT unioninput.*
FROM (
FROM src select src.key, src.value WHERE src.key < 100
@@ -1235,7 +1235,7 @@ FROM (
CLUSTER BY unioninput.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-57_574_7730562904986793585/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-03-50_185_4510032647777990341/-mr-10000
0 val_0
0 val_0
0 val_0
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/filter_join_breaktask.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/filter_join_breaktask.q.out?rev=980659&r1=980658&r2=980659&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/filter_join_breaktask.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/filter_join_breaktask.q.out Fri Jul 30 06:40:04 2010
@@ -98,9 +98,9 @@ STAGE PLANS:
type: string
Needs Tagging: true
Path -> Alias:
- pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/filter_join_breaktask/ds=2008-04-08 [f, m]
+ pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/filter_join_breaktask/ds=2008-04-08 [f, m]
Path -> Partition:
- pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/filter_join_breaktask/ds=2008-04-08
+ pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/filter_join_breaktask/ds=2008-04-08
Partition
base file name: ds=2008-04-08
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -113,13 +113,13 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/filter_join_breaktask
+ location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/filter_join_breaktask
name filter_join_breaktask
partition_columns ds
serialization.ddl struct filter_join_breaktask { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1280083899
+ transient_lastDdlTime 1280427201
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -130,13 +130,13 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/filter_join_breaktask
+ location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/filter_join_breaktask
name filter_join_breaktask
partition_columns ds
serialization.ddl struct filter_join_breaktask { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1280083899
+ transient_lastDdlTime 1280427201
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: filter_join_breaktask
name: filter_join_breaktask
@@ -148,32 +148,32 @@ STAGE PLANS:
0 {VALUE._col0}
1 {VALUE._col1} {VALUE._col2}
handleSkewJoin: false
- outputColumnNames: _col0, _col4, _col5
+ outputColumnNames: _col0, _col6, _col7
Filter Operator
isSamplingPred: false
predicate:
- expr: (_col5 = '2008-04-08')
+ expr: (_col7 = '2008-04-08')
type: boolean
Filter Operator
isSamplingPred: false
predicate:
- expr: _col4 is not null
+ expr: _col6 is not null
type: boolean
Filter Operator
isSamplingPred: false
predicate:
- expr: (_col4 <> '')
+ expr: (_col6 <> '')
type: boolean
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/tmp/jssarma/hive_2010-07-25_11-51-42_736_7752313177772027899/-mr-10002
+ directory: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-13-27_145_4349278496196246379/-mr-10002
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
properties:
- columns _col0,_col4
+ columns _col0,_col6
columns.types int,string
escape.delim \
TotalFiles: 1
@@ -185,11 +185,11 @@ STAGE PLANS:
$INTNAME
Reduce Output Operator
key expressions:
- expr: _col4
+ expr: _col6
type: string
sort order: +
Map-reduce partition columns:
- expr: _col4
+ expr: _col6
type: string
tag: 0
value expressions:
@@ -222,26 +222,26 @@ STAGE PLANS:
type: string
Needs Tagging: true
Path -> Alias:
- file:/tmp/jssarma/hive_2010-07-25_11-51-42_736_7752313177772027899/-mr-10002 [$INTNAME]
- pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/filter_join_breaktask/ds=2008-04-08 [g]
+ file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-13-27_145_4349278496196246379/-mr-10002 [$INTNAME]
+ pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/filter_join_breaktask/ds=2008-04-08 [g]
Path -> Partition:
- file:/tmp/jssarma/hive_2010-07-25_11-51-42_736_7752313177772027899/-mr-10002
+ file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-13-27_145_4349278496196246379/-mr-10002
Partition
base file name: -mr-10002
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
properties:
- columns _col0,_col4
+ columns _col0,_col6
columns.types int,string
escape.delim \
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
properties:
- columns _col0,_col4
+ columns _col0,_col6
columns.types int,string
escape.delim \
- pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/filter_join_breaktask/ds=2008-04-08
+ pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/filter_join_breaktask/ds=2008-04-08
Partition
base file name: ds=2008-04-08
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -254,13 +254,13 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/filter_join_breaktask
+ location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/filter_join_breaktask
name filter_join_breaktask
partition_columns ds
serialization.ddl struct filter_join_breaktask { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1280083899
+ transient_lastDdlTime 1280427201
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -271,13 +271,13 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/filter_join_breaktask
+ location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/filter_join_breaktask
name filter_join_breaktask
partition_columns ds
serialization.ddl struct filter_join_breaktask { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1280083899
+ transient_lastDdlTime 1280427201
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: filter_join_breaktask
name: filter_join_breaktask
@@ -289,18 +289,18 @@ STAGE PLANS:
0 {VALUE._col0}
1 {VALUE._col1}
handleSkewJoin: false
- outputColumnNames: _col0, _col7
+ outputColumnNames: _col0, _col11
Select Operator
expressions:
expr: _col0
type: int
- expr: _col7
+ expr: _col11
type: string
outputColumnNames: _col0, _col1
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/tmp/jssarma/hive_2010-07-25_11-51-42_736_7752313177772027899/-ext-10001
+ directory: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-13-27_145_4349278496196246379/-ext-10001
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -322,13 +322,13 @@ FROM filter_join_breaktask f JOIN filter
JOIN filter_join_breaktask g ON(g.value = m.value AND g.ds='2008-04-08' AND m.ds='2008-04-08' AND m.value is not null AND m.value !='')
PREHOOK: type: QUERY
PREHOOK: Input: default@filter_join_breaktask@ds=2008-04-08
-PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-25_11-51-42_928_8130623629772724799/-mr-10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-13-30_238_1373557843480629621/-mr-10000
POSTHOOK: query: SELECT f.key, g.value
FROM filter_join_breaktask f JOIN filter_join_breaktask m ON( f.key = m.key AND f.ds='2008-04-08' AND m.ds='2008-04-08' AND f.key is not null)
JOIN filter_join_breaktask g ON(g.value = m.value AND g.ds='2008-04-08' AND m.ds='2008-04-08' AND m.value is not null AND m.value !='')
POSTHOOK: type: QUERY
POSTHOOK: Input: default@filter_join_breaktask@ds=2008-04-08
-POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-25_11-51-42_928_8130623629772724799/-mr-10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_11-13-30_238_1373557843480629621/-mr-10000
POSTHOOK: Lineage: filter_join_breaktask PARTITION(ds=2008-04-08).key EXPRESSION [(src1)src1.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: filter_join_breaktask PARTITION(ds=2008-04-08).value SIMPLE [(src1)src1.FieldSchema(name:value, type:string, comment:default), ]
146 val_146