You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by jv...@apache.org on 2011/07/13 18:49:42 UTC
svn commit: r1146129 [16/41] - in /hive/trunk:
common/src/java/org/apache/hadoop/hive/conf/
contrib/src/test/results/clientpositive/ hbase-handler/src/test/results/
ql/src/java/org/apache/hadoop/hive/ql/exec/
ql/src/java/org/apache/hadoop/hive/ql/ppd/ ...
Modified: hive/trunk/ql/src/test/results/clientpositive/sample6.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/sample6.q.out?rev=1146129&r1=1146128&r2=1146129&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/sample6.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/sample6.q.out Wed Jul 13 16:49:22 2011
@@ -33,55 +33,50 @@ STAGE PLANS:
alias: s
GatherStats: false
Filter Operator
- isSamplingPred: false
+ isSamplingPred: true
predicate:
expr: (((hash(key) & 2147483647) % 4) = 0)
type: boolean
- Filter Operator
- isSamplingPred: true
- predicate:
- expr: (((hash(key) & 2147483647) % 4) = 0)
- type: boolean
- Select Operator
- expressions:
- expr: key
- type: int
- expr: value
- type: string
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- GlobalTableId: 1
- directory: pfile:/data/users/sdong/www/open-source-hive1/build/ql/scratchdir/hive_2011-04-19_19-54-27_627_3309354126941173289/-ext-10002
- NumFilesPerFileSink: 1
- Stats Publishing Key Prefix: pfile:/data/users/sdong/www/open-source-hive1/build/ql/scratchdir/hive_2011-04-19_19-54-27_627_3309354126941173289/-ext-10000/
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- properties:
- bucket_count -1
- columns key,value
- columns.types int:string
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/dest1
- name default.dest1
- serialization.ddl struct dest1 { i32 key, string value}
- serialization.format 1
- serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303268067
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.dest1
- TotalFiles: 1
- GatherStats: true
- MultiFileSpray: false
+ Select Operator
+ expressions:
+ expr: key
+ type: int
+ expr: value
+ type: string
+ outputColumnNames: _col0, _col1
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ directory: pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2011-07-11_02-58-00_018_458728813403565509/-ext-10002
+ NumFilesPerFileSink: 1
+ Stats Publishing Key Prefix: pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2011-07-11_02-58-00_018_458728813403565509/-ext-10000/
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ bucket_count -1
+ columns key,value
+ columns.types int:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1
+ name default.dest1
+ serialization.ddl struct dest1 { i32 key, string value}
+ serialization.format 1
+ serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ transient_lastDdlTime 1310378280
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.dest1
+ TotalFiles: 1
+ GatherStats: true
+ MultiFileSpray: false
Needs Tagging: false
Path -> Alias:
- pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt [s]
+ pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket [s]
Path -> Partition:
- pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt
+ pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket
Partition
- base file name: srcbucket0.txt
+ base file name: srcbucket
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
@@ -91,12 +86,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket
name default.srcbucket
serialization.ddl struct srcbucket { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303260542
+ transient_lastDdlTime 1310371847
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -108,12 +103,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket
name default.srcbucket
serialization.ddl struct srcbucket { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303260542
+ transient_lastDdlTime 1310371847
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket
name: default.srcbucket
@@ -125,14 +120,14 @@ STAGE PLANS:
Move Operator
files:
hdfs directory: true
- source: pfile:/data/users/sdong/www/open-source-hive1/build/ql/scratchdir/hive_2011-04-19_19-54-27_627_3309354126941173289/-ext-10002
- destination: pfile:/data/users/sdong/www/open-source-hive1/build/ql/scratchdir/hive_2011-04-19_19-54-27_627_3309354126941173289/-ext-10000
+ source: pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2011-07-11_02-58-00_018_458728813403565509/-ext-10002
+ destination: pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2011-07-11_02-58-00_018_458728813403565509/-ext-10000
Stage: Stage-0
Move Operator
tables:
replace: true
- source: pfile:/data/users/sdong/www/open-source-hive1/build/ql/scratchdir/hive_2011-04-19_19-54-27_627_3309354126941173289/-ext-10000
+ source: pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2011-07-11_02-58-00_018_458728813403565509/-ext-10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -142,28 +137,28 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/dest1
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1
name default.dest1
serialization.ddl struct dest1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303268067
+ transient_lastDdlTime 1310378280
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1
- tmp directory: pfile:/data/users/sdong/www/open-source-hive1/build/ql/scratchdir/hive_2011-04-19_19-54-27_627_3309354126941173289/-ext-10001
+ tmp directory: pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2011-07-11_02-58-00_018_458728813403565509/-ext-10001
Stage: Stage-2
Stats-Aggr Operator
- Stats Aggregation Key Prefix: pfile:/data/users/sdong/www/open-source-hive1/build/ql/scratchdir/hive_2011-04-19_19-54-27_627_3309354126941173289/-ext-10000/
+ Stats Aggregation Key Prefix: pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2011-07-11_02-58-00_018_458728813403565509/-ext-10000/
Stage: Stage-3
Map Reduce
Alias -> Map Operator Tree:
- pfile:/data/users/sdong/www/open-source-hive1/build/ql/scratchdir/hive_2011-04-19_19-54-27_627_3309354126941173289/-ext-10002
+ pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2011-07-11_02-58-00_018_458728813403565509/-ext-10002
File Output Operator
compressed: false
GlobalTableId: 0
- directory: pfile:/data/users/sdong/www/open-source-hive1/build/ql/scratchdir/hive_2011-04-19_19-54-27_627_3309354126941173289/-ext-10000
+ directory: pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2011-07-11_02-58-00_018_458728813403565509/-ext-10000
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -174,12 +169,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/dest1
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1
name default.dest1
serialization.ddl struct dest1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303268067
+ transient_lastDdlTime 1310378280
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1
TotalFiles: 1
@@ -187,9 +182,9 @@ STAGE PLANS:
MultiFileSpray: false
Needs Tagging: false
Path -> Alias:
- pfile:/data/users/sdong/www/open-source-hive1/build/ql/scratchdir/hive_2011-04-19_19-54-27_627_3309354126941173289/-ext-10002 [pfile:/data/users/sdong/www/open-source-hive1/build/ql/scratchdir/hive_2011-04-19_19-54-27_627_3309354126941173289/-ext-10002]
+ pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2011-07-11_02-58-00_018_458728813403565509/-ext-10002 [pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2011-07-11_02-58-00_018_458728813403565509/-ext-10002]
Path -> Partition:
- pfile:/data/users/sdong/www/open-source-hive1/build/ql/scratchdir/hive_2011-04-19_19-54-27_627_3309354126941173289/-ext-10002
+ pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2011-07-11_02-58-00_018_458728813403565509/-ext-10002
Partition
base file name: -ext-10002
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -200,12 +195,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/dest1
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1
name default.dest1
serialization.ddl struct dest1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303268067
+ transient_lastDdlTime 1310378280
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -216,12 +211,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/dest1
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1
name default.dest1
serialization.ddl struct dest1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303268067
+ transient_lastDdlTime 1310378280
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1
name: default.dest1
@@ -242,11 +237,11 @@ POSTHOOK: Lineage: dest1.value SIMPLE [(
PREHOOK: query: SELECT dest1.* FROM dest1
PREHOOK: type: QUERY
PREHOOK: Input: default@dest1
-PREHOOK: Output: file:/tmp/sdong/hive_2011-04-19_19-54-33_389_7760677817749293223/-mr-10000
+PREHOOK: Output: file:/tmp/amarsri/hive_2011-07-11_02-58-03_687_2062247448126517612/-mr-10000
POSTHOOK: query: SELECT dest1.* FROM dest1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest1
-POSTHOOK: Output: file:/tmp/sdong/hive_2011-04-19_19-54-33_389_7760677817749293223/-mr-10000
+POSTHOOK: Output: file:/tmp/amarsri/hive_2011-07-11_02-58-03_687_2062247448126517612/-mr-10000
POSTHOOK: Lineage: dest1.key SIMPLE [(srcbucket)s.FieldSchema(name:key, type:int, comment:null), ]
POSTHOOK: Lineage: dest1.value SIMPLE [(srcbucket)s.FieldSchema(name:value, type:string, comment:null), ]
468 val_469
@@ -522,42 +517,37 @@ STAGE PLANS:
alias: s
GatherStats: false
Filter Operator
- isSamplingPred: false
+ isSamplingPred: true
predicate:
expr: (((hash(key) & 2147483647) % 4) = 3)
type: boolean
- Filter Operator
- isSamplingPred: true
- predicate:
- expr: (((hash(key) & 2147483647) % 4) = 3)
- type: boolean
- Select Operator
- expressions:
- expr: key
+ Select Operator
+ expressions:
+ expr: key
+ type: int
+ expr: value
+ type: string
+ outputColumnNames: _col0, _col1
+ Reduce Output Operator
+ key expressions:
+ expr: _col0
+ type: int
+ expr: _col1
+ type: string
+ sort order: ++
+ tag: -1
+ value expressions:
+ expr: _col0
type: int
- expr: value
+ expr: _col1
type: string
- outputColumnNames: _col0, _col1
- Reduce Output Operator
- key expressions:
- expr: _col0
- type: int
- expr: _col1
- type: string
- sort order: ++
- tag: -1
- value expressions:
- expr: _col0
- type: int
- expr: _col1
- type: string
Needs Tagging: false
Path -> Alias:
- pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket/srcbucket1.txt [s]
+ pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket [s]
Path -> Partition:
- pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket/srcbucket1.txt
+ pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket
Partition
- base file name: srcbucket1.txt
+ base file name: srcbucket
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
@@ -567,12 +557,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket
name default.srcbucket
serialization.ddl struct srcbucket { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303260542
+ transient_lastDdlTime 1310371847
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -584,12 +574,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket
name default.srcbucket
serialization.ddl struct srcbucket { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303260542
+ transient_lastDdlTime 1310371847
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket
name: default.srcbucket
@@ -598,9 +588,9 @@ STAGE PLANS:
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/tmp/sdong/hive_2011-04-19_19-54-33_827_389668384025414462/-ext-10001
+ directory: file:/tmp/amarsri/hive_2011-07-11_02-58-03_815_3440650522759974343/-ext-10001
NumFilesPerFileSink: 1
- Stats Publishing Key Prefix: file:/tmp/sdong/hive_2011-04-19_19-54-33_827_389668384025414462/-ext-10001/
+ Stats Publishing Key Prefix: file:/tmp/amarsri/hive_2011-07-11_02-58-03_815_3440650522759974343/-ext-10001/
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -621,12 +611,12 @@ PREHOOK: query: SELECT s.* FROM srcbucke
ORDER BY key, value
PREHOOK: type: QUERY
PREHOOK: Input: default@srcbucket
-PREHOOK: Output: file:/tmp/sdong/hive_2011-04-19_19-54-34_030_551214422770351063/-mr-10000
+PREHOOK: Output: file:/tmp/amarsri/hive_2011-07-11_02-58-03_854_6271750517263563878/-mr-10000
POSTHOOK: query: SELECT s.* FROM srcbucket TABLESAMPLE (BUCKET 4 OUT OF 4 on key) s
ORDER BY key, value
POSTHOOK: type: QUERY
POSTHOOK: Input: default@srcbucket
-POSTHOOK: Output: file:/tmp/sdong/hive_2011-04-19_19-54-34_030_551214422770351063/-mr-10000
+POSTHOOK: Output: file:/tmp/amarsri/hive_2011-07-11_02-58-03_854_6271750517263563878/-mr-10000
POSTHOOK: Lineage: dest1.key SIMPLE [(srcbucket)s.FieldSchema(name:key, type:int, comment:null), ]
POSTHOOK: Lineage: dest1.value SIMPLE [(srcbucket)s.FieldSchema(name:value, type:string, comment:null), ]
3 val_4
@@ -892,42 +882,37 @@ STAGE PLANS:
alias: s
GatherStats: false
Filter Operator
- isSamplingPred: false
+ isSamplingPred: true
predicate:
expr: (((hash(key) & 2147483647) % 2) = 0)
type: boolean
- Filter Operator
- isSamplingPred: true
- predicate:
- expr: (((hash(key) & 2147483647) % 2) = 0)
- type: boolean
- Select Operator
- expressions:
- expr: key
+ Select Operator
+ expressions:
+ expr: key
+ type: int
+ expr: value
+ type: string
+ outputColumnNames: _col0, _col1
+ Reduce Output Operator
+ key expressions:
+ expr: _col0
+ type: int
+ expr: _col1
+ type: string
+ sort order: ++
+ tag: -1
+ value expressions:
+ expr: _col0
type: int
- expr: value
+ expr: _col1
type: string
- outputColumnNames: _col0, _col1
- Reduce Output Operator
- key expressions:
- expr: _col0
- type: int
- expr: _col1
- type: string
- sort order: ++
- tag: -1
- value expressions:
- expr: _col0
- type: int
- expr: _col1
- type: string
Needs Tagging: false
Path -> Alias:
- pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt [s]
+ pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket [s]
Path -> Partition:
- pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt
+ pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket
Partition
- base file name: srcbucket0.txt
+ base file name: srcbucket
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
@@ -937,12 +922,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket
name default.srcbucket
serialization.ddl struct srcbucket { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303260542
+ transient_lastDdlTime 1310371847
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -954,12 +939,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket
name default.srcbucket
serialization.ddl struct srcbucket { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303260542
+ transient_lastDdlTime 1310371847
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket
name: default.srcbucket
@@ -968,9 +953,9 @@ STAGE PLANS:
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/tmp/sdong/hive_2011-04-19_19-54-37_207_200034802818981990/-ext-10001
+ directory: file:/tmp/amarsri/hive_2011-07-11_02-58-06_669_1410317577849689409/-ext-10001
NumFilesPerFileSink: 1
- Stats Publishing Key Prefix: file:/tmp/sdong/hive_2011-04-19_19-54-37_207_200034802818981990/-ext-10001/
+ Stats Publishing Key Prefix: file:/tmp/amarsri/hive_2011-07-11_02-58-06_669_1410317577849689409/-ext-10001/
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -991,12 +976,12 @@ PREHOOK: query: SELECT s.* FROM srcbucke
ORDER BY key, value
PREHOOK: type: QUERY
PREHOOK: Input: default@srcbucket
-PREHOOK: Output: file:/tmp/sdong/hive_2011-04-19_19-54-37_421_4651239266027596073/-mr-10000
+PREHOOK: Output: file:/tmp/amarsri/hive_2011-07-11_02-58-06_708_7756573301727009933/-mr-10000
POSTHOOK: query: SELECT s.* FROM srcbucket TABLESAMPLE (BUCKET 1 OUT OF 2 on key) s
ORDER BY key, value
POSTHOOK: type: QUERY
POSTHOOK: Input: default@srcbucket
-POSTHOOK: Output: file:/tmp/sdong/hive_2011-04-19_19-54-37_421_4651239266027596073/-mr-10000
+POSTHOOK: Output: file:/tmp/amarsri/hive_2011-07-11_02-58-06_708_7756573301727009933/-mr-10000
POSTHOOK: Lineage: dest1.key SIMPLE [(srcbucket)s.FieldSchema(name:key, type:int, comment:null), ]
POSTHOOK: Lineage: dest1.value SIMPLE [(srcbucket)s.FieldSchema(name:value, type:string, comment:null), ]
0 val_0
@@ -1516,40 +1501,35 @@ STAGE PLANS:
alias: s
GatherStats: false
Filter Operator
- isSamplingPred: false
+ isSamplingPred: true
predicate:
expr: (((hash(key) & 2147483647) % 3) = 0)
type: boolean
- Filter Operator
- isSamplingPred: true
- predicate:
- expr: (((hash(key) & 2147483647) % 3) = 0)
- type: boolean
- Select Operator
- expressions:
- expr: key
+ Select Operator
+ expressions:
+ expr: key
+ type: int
+ expr: value
+ type: string
+ outputColumnNames: _col0, _col1
+ Reduce Output Operator
+ key expressions:
+ expr: _col0
type: int
- expr: value
+ expr: _col1
+ type: string
+ sort order: ++
+ tag: -1
+ value expressions:
+ expr: _col0
+ type: int
+ expr: _col1
type: string
- outputColumnNames: _col0, _col1
- Reduce Output Operator
- key expressions:
- expr: _col0
- type: int
- expr: _col1
- type: string
- sort order: ++
- tag: -1
- value expressions:
- expr: _col0
- type: int
- expr: _col1
- type: string
Needs Tagging: false
Path -> Alias:
- pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket [s]
+ pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket [s]
Path -> Partition:
- pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket
+ pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket
Partition
base file name: srcbucket
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -1561,12 +1541,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket
name default.srcbucket
serialization.ddl struct srcbucket { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303260542
+ transient_lastDdlTime 1310371847
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -1578,12 +1558,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket
name default.srcbucket
serialization.ddl struct srcbucket { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303260542
+ transient_lastDdlTime 1310371847
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket
name: default.srcbucket
@@ -1592,9 +1572,9 @@ STAGE PLANS:
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/tmp/sdong/hive_2011-04-19_19-54-40_626_8344413354265807082/-ext-10001
+ directory: file:/tmp/amarsri/hive_2011-07-11_02-58-09_360_6989967716610636074/-ext-10001
NumFilesPerFileSink: 1
- Stats Publishing Key Prefix: file:/tmp/sdong/hive_2011-04-19_19-54-40_626_8344413354265807082/-ext-10001/
+ Stats Publishing Key Prefix: file:/tmp/amarsri/hive_2011-07-11_02-58-09_360_6989967716610636074/-ext-10001/
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -1615,12 +1595,12 @@ PREHOOK: query: SELECT s.* FROM srcbucke
ORDER BY key, value
PREHOOK: type: QUERY
PREHOOK: Input: default@srcbucket
-PREHOOK: Output: file:/tmp/sdong/hive_2011-04-19_19-54-40_714_2969909884270153149/-mr-10000
+PREHOOK: Output: file:/tmp/amarsri/hive_2011-07-11_02-58-09_405_8746850347504403733/-mr-10000
POSTHOOK: query: SELECT s.* FROM srcbucket TABLESAMPLE (BUCKET 1 OUT OF 3 on key) s
ORDER BY key, value
POSTHOOK: type: QUERY
POSTHOOK: Input: default@srcbucket
-POSTHOOK: Output: file:/tmp/sdong/hive_2011-04-19_19-54-40_714_2969909884270153149/-mr-10000
+POSTHOOK: Output: file:/tmp/amarsri/hive_2011-07-11_02-58-09_405_8746850347504403733/-mr-10000
POSTHOOK: Lineage: dest1.key SIMPLE [(srcbucket)s.FieldSchema(name:key, type:int, comment:null), ]
POSTHOOK: Lineage: dest1.value SIMPLE [(srcbucket)s.FieldSchema(name:value, type:string, comment:null), ]
0 val_0
@@ -1983,40 +1963,35 @@ STAGE PLANS:
alias: s
GatherStats: false
Filter Operator
- isSamplingPred: false
+ isSamplingPred: true
predicate:
expr: (((hash(key) & 2147483647) % 3) = 1)
type: boolean
- Filter Operator
- isSamplingPred: true
- predicate:
- expr: (((hash(key) & 2147483647) % 3) = 1)
- type: boolean
- Select Operator
- expressions:
- expr: key
+ Select Operator
+ expressions:
+ expr: key
+ type: int
+ expr: value
+ type: string
+ outputColumnNames: _col0, _col1
+ Reduce Output Operator
+ key expressions:
+ expr: _col0
type: int
- expr: value
+ expr: _col1
+ type: string
+ sort order: ++
+ tag: -1
+ value expressions:
+ expr: _col0
+ type: int
+ expr: _col1
type: string
- outputColumnNames: _col0, _col1
- Reduce Output Operator
- key expressions:
- expr: _col0
- type: int
- expr: _col1
- type: string
- sort order: ++
- tag: -1
- value expressions:
- expr: _col0
- type: int
- expr: _col1
- type: string
Needs Tagging: false
Path -> Alias:
- pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket [s]
+ pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket [s]
Path -> Partition:
- pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket
+ pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket
Partition
base file name: srcbucket
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -2028,12 +2003,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket
name default.srcbucket
serialization.ddl struct srcbucket { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303260542
+ transient_lastDdlTime 1310371847
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -2045,12 +2020,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket
name default.srcbucket
serialization.ddl struct srcbucket { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303260542
+ transient_lastDdlTime 1310371847
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket
name: default.srcbucket
@@ -2059,9 +2034,9 @@ STAGE PLANS:
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/tmp/sdong/hive_2011-04-19_19-54-43_892_7813003909078968435/-ext-10001
+ directory: file:/tmp/amarsri/hive_2011-07-11_02-58-12_091_1241249343366878880/-ext-10001
NumFilesPerFileSink: 1
- Stats Publishing Key Prefix: file:/tmp/sdong/hive_2011-04-19_19-54-43_892_7813003909078968435/-ext-10001/
+ Stats Publishing Key Prefix: file:/tmp/amarsri/hive_2011-07-11_02-58-12_091_1241249343366878880/-ext-10001/
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -2082,12 +2057,12 @@ PREHOOK: query: SELECT s.* FROM srcbucke
ORDER BY key, value
PREHOOK: type: QUERY
PREHOOK: Input: default@srcbucket
-PREHOOK: Output: file:/tmp/sdong/hive_2011-04-19_19-54-43_980_1056408795977903895/-mr-10000
+PREHOOK: Output: file:/tmp/amarsri/hive_2011-07-11_02-58-12_130_7467195593406386569/-mr-10000
POSTHOOK: query: SELECT s.* FROM srcbucket TABLESAMPLE (BUCKET 2 OUT OF 3 on key) s
ORDER BY key, value
POSTHOOK: type: QUERY
POSTHOOK: Input: default@srcbucket
-POSTHOOK: Output: file:/tmp/sdong/hive_2011-04-19_19-54-43_980_1056408795977903895/-mr-10000
+POSTHOOK: Output: file:/tmp/amarsri/hive_2011-07-11_02-58-12_130_7467195593406386569/-mr-10000
POSTHOOK: Lineage: dest1.key SIMPLE [(srcbucket)s.FieldSchema(name:key, type:int, comment:null), ]
POSTHOOK: Lineage: dest1.value SIMPLE [(srcbucket)s.FieldSchema(name:value, type:string, comment:null), ]
1 val_2
@@ -2436,43 +2411,37 @@ STAGE PLANS:
alias: s
GatherStats: false
Filter Operator
- isSamplingPred: false
+ isSamplingPred: true
predicate:
expr: (((hash(key) & 2147483647) % 2) = 0)
type: boolean
- Filter Operator
- isSamplingPred: true
- predicate:
- expr: (((hash(key) & 2147483647) % 2) = 0)
- type: boolean
- Select Operator
- expressions:
- expr: key
+ Select Operator
+ expressions:
+ expr: key
+ type: int
+ expr: value
+ type: string
+ outputColumnNames: _col0, _col1
+ Reduce Output Operator
+ key expressions:
+ expr: _col0
type: int
- expr: value
+ expr: _col1
+ type: string
+ sort order: ++
+ tag: -1
+ value expressions:
+ expr: _col0
+ type: int
+ expr: _col1
type: string
- outputColumnNames: _col0, _col1
- Reduce Output Operator
- key expressions:
- expr: _col0
- type: int
- expr: _col1
- type: string
- sort order: ++
- tag: -1
- value expressions:
- expr: _col0
- type: int
- expr: _col1
- type: string
Needs Tagging: false
Path -> Alias:
- pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket2/srcbucket20.txt [s]
- pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket2/srcbucket22.txt [s]
+ pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket2 [s]
Path -> Partition:
- pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket2/srcbucket20.txt
+ pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket2
Partition
- base file name: srcbucket20.txt
+ base file name: srcbucket2
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
@@ -2482,12 +2451,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket2
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket2
name default.srcbucket2
serialization.ddl struct srcbucket2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303260544
+ transient_lastDdlTime 1310371848
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -2499,50 +2468,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket2
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket2
name default.srcbucket2
serialization.ddl struct srcbucket2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303260544
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.srcbucket2
- name: default.srcbucket2
- pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket2/srcbucket22.txt
- Partition
- base file name: srcbucket22.txt
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- properties:
- bucket_count 4
- bucket_field_name key
- columns key,value
- columns.types int:string
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket2
- name default.srcbucket2
- serialization.ddl struct srcbucket2 { i32 key, string value}
- serialization.format 1
- serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303260544
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- properties:
- bucket_count 4
- bucket_field_name key
- columns key,value
- columns.types int:string
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket2
- name default.srcbucket2
- serialization.ddl struct srcbucket2 { i32 key, string value}
- serialization.format 1
- serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303260544
+ transient_lastDdlTime 1310371848
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket2
name: default.srcbucket2
@@ -2551,9 +2482,9 @@ STAGE PLANS:
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/tmp/sdong/hive_2011-04-19_19-54-47_115_1778465985574852574/-ext-10001
+ directory: file:/tmp/amarsri/hive_2011-07-11_02-58-14_778_4777537004727741553/-ext-10001
NumFilesPerFileSink: 1
- Stats Publishing Key Prefix: file:/tmp/sdong/hive_2011-04-19_19-54-47_115_1778465985574852574/-ext-10001/
+ Stats Publishing Key Prefix: file:/tmp/amarsri/hive_2011-07-11_02-58-14_778_4777537004727741553/-ext-10001/
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -2574,12 +2505,12 @@ PREHOOK: query: SELECT s.* FROM srcbucke
ORDER BY key, value
PREHOOK: type: QUERY
PREHOOK: Input: default@srcbucket2
-PREHOOK: Output: file:/tmp/sdong/hive_2011-04-19_19-54-47_633_8309063970411271815/-mr-10000
+PREHOOK: Output: file:/tmp/amarsri/hive_2011-07-11_02-58-14_817_7365496467903983867/-mr-10000
POSTHOOK: query: SELECT s.* FROM srcbucket2 TABLESAMPLE (BUCKET 1 OUT OF 2 on key) s
ORDER BY key, value
POSTHOOK: type: QUERY
POSTHOOK: Input: default@srcbucket2
-POSTHOOK: Output: file:/tmp/sdong/hive_2011-04-19_19-54-47_633_8309063970411271815/-mr-10000
+POSTHOOK: Output: file:/tmp/amarsri/hive_2011-07-11_02-58-14_817_7365496467903983867/-mr-10000
POSTHOOK: Lineage: dest1.key SIMPLE [(srcbucket)s.FieldSchema(name:key, type:int, comment:null), ]
POSTHOOK: Lineage: dest1.value SIMPLE [(srcbucket)s.FieldSchema(name:value, type:string, comment:null), ]
0 val_0
@@ -2588,26 +2519,62 @@ POSTHOOK: Lineage: dest1.value SIMPLE [(
2 val_2
4 val_4
8 val_8
+10 val_10
+12 val_12
+12 val_12
+18 val_18
+18 val_18
20 val_20
24 val_24
24 val_24
26 val_26
26 val_26
28 val_28
+30 val_30
+34 val_34
42 val_42
42 val_42
44 val_44
+54 val_54
+58 val_58
+58 val_58
64 val_64
66 val_66
+70 val_70
+70 val_70
+70 val_70
+72 val_72
+72 val_72
+74 val_74
+76 val_76
+76 val_76
+78 val_78
80 val_80
82 val_82
84 val_84
84 val_84
86 val_86
+90 val_90
+90 val_90
+90 val_90
+92 val_92
+96 val_96
+98 val_98
+98 val_98
+100 val_100
+100 val_100
+104 val_104
+104 val_104
114 val_114
116 val_116
118 val_118
118 val_118
+120 val_120
+120 val_120
+126 val_126
+128 val_128
+128 val_128
+128 val_128
134 val_134
134 val_134
136 val_136
@@ -2615,11 +2582,19 @@ POSTHOOK: Lineage: dest1.value SIMPLE [(
138 val_138
138 val_138
138 val_138
+146 val_146
+146 val_146
150 val_150
152 val_152
152 val_152
156 val_156
158 val_158
+160 val_160
+162 val_162
+164 val_164
+164 val_164
+166 val_166
+168 val_168
170 val_170
172 val_172
172 val_172
@@ -2628,6 +2603,8 @@ POSTHOOK: Lineage: dest1.value SIMPLE [(
176 val_176
176 val_176
178 val_178
+180 val_180
+186 val_186
190 val_190
192 val_192
194 val_194
@@ -2638,18 +2615,38 @@ POSTHOOK: Lineage: dest1.value SIMPLE [(
208 val_208
208 val_208
208 val_208
+214 val_214
+216 val_216
+216 val_216
+218 val_218
222 val_222
224 val_224
224 val_224
226 val_226
228 val_228
+230 val_230
+230 val_230
+230 val_230
+230 val_230
+230 val_230
+238 val_238
+238 val_238
242 val_242
242 val_242
244 val_244
248 val_248
+252 val_252
+256 val_256
+256 val_256
+258 val_258
260 val_260
262 val_262
266 val_266
+272 val_272
+272 val_272
+274 val_274
+278 val_278
+278 val_278
280 val_280
280 val_280
282 val_282
@@ -2658,6 +2655,14 @@ POSTHOOK: Lineage: dest1.value SIMPLE [(
286 val_286
288 val_288
288 val_288
+292 val_292
+296 val_296
+298 val_298
+298 val_298
+298 val_298
+302 val_302
+306 val_306
+308 val_308
310 val_310
316 val_316
316 val_316
@@ -2665,12 +2670,34 @@ POSTHOOK: Lineage: dest1.value SIMPLE [(
318 val_318
318 val_318
318 val_318
+322 val_322
+322 val_322
332 val_332
336 val_336
338 val_338
+342 val_342
+342 val_342
+344 val_344
+344 val_344
+348 val_348
+348 val_348
+348 val_348
+348 val_348
+348 val_348
356 val_356
+360 val_360
+362 val_362
+364 val_364
+366 val_366
+368 val_368
374 val_374
378 val_378
+382 val_382
+382 val_382
+384 val_384
+384 val_384
+384 val_384
+386 val_386
392 val_392
394 val_394
396 val_396
@@ -2684,11 +2711,28 @@ POSTHOOK: Lineage: dest1.value SIMPLE [(
406 val_406
406 val_406
406 val_406
+414 val_414
+414 val_414
+418 val_418
424 val_424
424 val_424
+430 val_430
+430 val_430
+430 val_430
+432 val_432
+436 val_436
+438 val_438
+438 val_438
+438 val_438
444 val_444
446 val_446
448 val_448
+452 val_452
+454 val_454
+454 val_454
+454 val_454
+458 val_458
+458 val_458
460 val_460
462 val_462
462 val_462
@@ -2699,11 +2743,23 @@ POSTHOOK: Lineage: dest1.value SIMPLE [(
468 val_468
468 val_468
468 val_468
+470 val_470
+472 val_472
+478 val_478
+478 val_478
480 val_480
480 val_480
480 val_480
482 val_482
484 val_484
+490 val_490
+492 val_492
+492 val_492
+494 val_494
+496 val_496
+498 val_498
+498 val_498
+498 val_498
PREHOOK: query: EXPLAIN EXTENDED SELECT s.* FROM srcbucket2 TABLESAMPLE (BUCKET 2 OUT OF 4 on key) s
ORDER BY key, value
PREHOOK: type: QUERY
@@ -2728,42 +2784,37 @@ STAGE PLANS:
alias: s
GatherStats: false
Filter Operator
- isSamplingPred: false
+ isSamplingPred: true
predicate:
expr: (((hash(key) & 2147483647) % 4) = 1)
type: boolean
- Filter Operator
- isSamplingPred: true
- predicate:
- expr: (((hash(key) & 2147483647) % 4) = 1)
- type: boolean
- Select Operator
- expressions:
- expr: key
+ Select Operator
+ expressions:
+ expr: key
+ type: int
+ expr: value
+ type: string
+ outputColumnNames: _col0, _col1
+ Reduce Output Operator
+ key expressions:
+ expr: _col0
type: int
- expr: value
+ expr: _col1
+ type: string
+ sort order: ++
+ tag: -1
+ value expressions:
+ expr: _col0
+ type: int
+ expr: _col1
type: string
- outputColumnNames: _col0, _col1
- Reduce Output Operator
- key expressions:
- expr: _col0
- type: int
- expr: _col1
- type: string
- sort order: ++
- tag: -1
- value expressions:
- expr: _col0
- type: int
- expr: _col1
- type: string
Needs Tagging: false
Path -> Alias:
- pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket2/srcbucket21.txt [s]
+ pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket2 [s]
Path -> Partition:
- pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket2/srcbucket21.txt
+ pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket2
Partition
- base file name: srcbucket21.txt
+ base file name: srcbucket2
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
@@ -2773,12 +2824,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket2
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket2
name default.srcbucket2
serialization.ddl struct srcbucket2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303260544
+ transient_lastDdlTime 1310371848
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -2790,12 +2841,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket2
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket2
name default.srcbucket2
serialization.ddl struct srcbucket2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303260544
+ transient_lastDdlTime 1310371848
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket2
name: default.srcbucket2
@@ -2804,9 +2855,9 @@ STAGE PLANS:
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/tmp/sdong/hive_2011-04-19_19-54-51_203_3981912443934260367/-ext-10001
+ directory: file:/tmp/amarsri/hive_2011-07-11_02-58-17_484_3834914689711003679/-ext-10001
NumFilesPerFileSink: 1
- Stats Publishing Key Prefix: file:/tmp/sdong/hive_2011-04-19_19-54-51_203_3981912443934260367/-ext-10001/
+ Stats Publishing Key Prefix: file:/tmp/amarsri/hive_2011-07-11_02-58-17_484_3834914689711003679/-ext-10001/
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -2827,34 +2878,115 @@ PREHOOK: query: SELECT s.* FROM srcbucke
ORDER BY key, value
PREHOOK: type: QUERY
PREHOOK: Input: default@srcbucket2
-PREHOOK: Output: file:/tmp/sdong/hive_2011-04-19_19-54-51_536_2222098349658422634/-mr-10000
+PREHOOK: Output: file:/tmp/amarsri/hive_2011-07-11_02-58-17_524_6116746546442362470/-mr-10000
POSTHOOK: query: SELECT s.* FROM srcbucket2 TABLESAMPLE (BUCKET 2 OUT OF 4 on key) s
ORDER BY key, value
POSTHOOK: type: QUERY
POSTHOOK: Input: default@srcbucket2
-POSTHOOK: Output: file:/tmp/sdong/hive_2011-04-19_19-54-51_536_2222098349658422634/-mr-10000
+POSTHOOK: Output: file:/tmp/amarsri/hive_2011-07-11_02-58-17_524_6116746546442362470/-mr-10000
POSTHOOK: Lineage: dest1.key SIMPLE [(srcbucket)s.FieldSchema(name:key, type:int, comment:null), ]
POSTHOOK: Lineage: dest1.value SIMPLE [(srcbucket)s.FieldSchema(name:value, type:string, comment:null), ]
5 val_5
5 val_5
5 val_5
9 val_9
+17 val_17
+33 val_33
+37 val_37
+37 val_37
41 val_41
+53 val_53
+57 val_57
+65 val_65
+69 val_69
+77 val_77
85 val_85
+97 val_97
+97 val_97
+105 val_105
+113 val_113
+113 val_113
+125 val_125
+125 val_125
+129 val_129
+129 val_129
133 val_133
137 val_137
137 val_137
+145 val_145
+149 val_149
+149 val_149
+153 val_153
+157 val_157
+165 val_165
+165 val_165
+169 val_169
+169 val_169
+169 val_169
+169 val_169
177 val_177
+181 val_181
+189 val_189
+193 val_193
+193 val_193
+193 val_193
+197 val_197
+197 val_197
+201 val_201
+205 val_205
+205 val_205
+209 val_209
+209 val_209
+213 val_213
+213 val_213
+217 val_217
+217 val_217
221 val_221
221 val_221
229 val_229
229 val_229
+233 val_233
+233 val_233
+237 val_237
+237 val_237
+241 val_241
+249 val_249
+257 val_257
265 val_265
265 val_265
+273 val_273
+273 val_273
+273 val_273
+277 val_277
+277 val_277
+277 val_277
+277 val_277
+281 val_281
+281 val_281
+285 val_285
+289 val_289
+305 val_305
+309 val_309
+309 val_309
317 val_317
317 val_317
+321 val_321
+321 val_321
+325 val_325
+325 val_325
+333 val_333
+333 val_333
+341 val_341
+345 val_345
353 val_353
353 val_353
+365 val_365
+369 val_369
+369 val_369
+369 val_369
+373 val_373
+377 val_377
+389 val_389
393 val_393
397 val_397
397 val_397
@@ -2866,13 +2998,32 @@ POSTHOOK: Lineage: dest1.value SIMPLE [(
409 val_409
409 val_409
409 val_409
+413 val_413
+413 val_413
+417 val_417
+417 val_417
+417 val_417
+421 val_421
+429 val_429
+429 val_429
+437 val_437
449 val_449
+453 val_453
+457 val_457
+469 val_469
+469 val_469
+469 val_469
+469 val_469
+469 val_469
+477 val_477
481 val_481
485 val_485
489 val_489
489 val_489
489 val_489
489 val_489
+493 val_493
+497 val_497
PREHOOK: query: CREATE TABLE empty_bucket (key int, value string) CLUSTERED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE empty_bucket (key int, value string) CLUSTERED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
@@ -2904,44 +3055,80 @@ STAGE PLANS:
alias: s
GatherStats: false
Filter Operator
- isSamplingPred: false
+ isSamplingPred: true
predicate:
expr: (((hash(key) & 2147483647) % 2) = 0)
type: boolean
- Filter Operator
- isSamplingPred: true
- predicate:
- expr: (((hash(key) & 2147483647) % 2) = 0)
- type: boolean
- Select Operator
- expressions:
- expr: key
+ Select Operator
+ expressions:
+ expr: key
+ type: int
+ expr: value
+ type: string
+ outputColumnNames: _col0, _col1
+ Reduce Output Operator
+ key expressions:
+ expr: _col0
type: int
- expr: value
+ expr: _col1
+ type: string
+ sort order: ++
+ tag: -1
+ value expressions:
+ expr: _col0
+ type: int
+ expr: _col1
type: string
- outputColumnNames: _col0, _col1
- Reduce Output Operator
- key expressions:
- expr: _col0
- type: int
- expr: _col1
- type: string
- sort order: ++
- tag: -1
- value expressions:
- expr: _col0
- type: int
- expr: _col1
- type: string
Needs Tagging: false
+ Path -> Alias:
+ pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/empty_bucket [s]
+ Path -> Partition:
+ pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/empty_bucket
+ Partition
+ base file name: empty_bucket
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ bucket_count 2
+ bucket_field_name key
+ columns key,value
+ columns.types int:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/empty_bucket
+ name default.empty_bucket
+ serialization.ddl struct empty_bucket { i32 key, string value}
+ serialization.format 1
+ serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ transient_lastDdlTime 1310378300
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ bucket_count 2
+ bucket_field_name key
+ columns key,value
+ columns.types int:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/empty_bucket
+ name default.empty_bucket
+ serialization.ddl struct empty_bucket { i32 key, string value}
+ serialization.format 1
+ serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ transient_lastDdlTime 1310378300
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.empty_bucket
+ name: default.empty_bucket
Reduce Operator Tree:
Extract
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/tmp/sdong/hive_2011-04-19_19-54-54_970_837635628196553945/-ext-10001
+ directory: file:/tmp/amarsri/hive_2011-07-11_02-58-20_219_7398883575090456577/-ext-10001
NumFilesPerFileSink: 1
- Stats Publishing Key Prefix: file:/tmp/sdong/hive_2011-04-19_19-54-54_970_837635628196553945/-ext-10001/
+ Stats Publishing Key Prefix: file:/tmp/amarsri/hive_2011-07-11_02-58-20_219_7398883575090456577/-ext-10001/
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -2962,11 +3149,11 @@ PREHOOK: query: SELECT s.* FROM empty_bu
ORDER BY key, value
PREHOOK: type: QUERY
PREHOOK: Input: default@empty_bucket
-PREHOOK: Output: file:/tmp/sdong/hive_2011-04-19_19-54-55_069_4633328979298722063/-mr-10000
+PREHOOK: Output: file:/tmp/amarsri/hive_2011-07-11_02-58-20_257_2520214403406970744/-mr-10000
POSTHOOK: query: SELECT s.* FROM empty_bucket TABLESAMPLE (BUCKET 1 OUT OF 2 on key) s
ORDER BY key, value
POSTHOOK: type: QUERY
POSTHOOK: Input: default@empty_bucket
-POSTHOOK: Output: file:/tmp/sdong/hive_2011-04-19_19-54-55_069_4633328979298722063/-mr-10000
+POSTHOOK: Output: file:/tmp/amarsri/hive_2011-07-11_02-58-20_257_2520214403406970744/-mr-10000
POSTHOOK: Lineage: dest1.key SIMPLE [(srcbucket)s.FieldSchema(name:key, type:int, comment:null), ]
POSTHOOK: Lineage: dest1.value SIMPLE [(srcbucket)s.FieldSchema(name:value, type:string, comment:null), ]
Modified: hive/trunk/ql/src/test/results/clientpositive/sample7.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out?rev=1146129&r1=1146128&r2=1146129&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/sample7.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/sample7.q.out Wed Jul 13 16:49:22 2011
@@ -37,51 +37,46 @@ STAGE PLANS:
Filter Operator
isSamplingPred: false
predicate:
- expr: ((((hash(key) & 2147483647) % 4) = 0) and (key > 100))
+ expr: (key > 100)
type: boolean
Filter Operator
isSamplingPred: true
predicate:
expr: (((hash(key) & 2147483647) % 4) = 0)
type: boolean
- Filter Operator
- isSamplingPred: false
- predicate:
- expr: (key > 100)
- type: boolean
- Select Operator
- expressions:
- expr: key
- type: int
- expr: value
- type: string
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- GlobalTableId: 1
- directory: pfile:/data/users/sdong/www/open-source-hive1/build/ql/scratchdir/hive_2011-04-19_19-54-58_518_7055483572049586597/-ext-10002
- NumFilesPerFileSink: 1
- Stats Publishing Key Prefix: pfile:/data/users/sdong/www/open-source-hive1/build/ql/scratchdir/hive_2011-04-19_19-54-58_518_7055483572049586597/-ext-10000/
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- properties:
- bucket_count -1
- columns key,value
- columns.types int:string
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/dest1
- name default.dest1
- serialization.ddl struct dest1 { i32 key, string value}
- serialization.format 1
- serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303268098
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.dest1
- TotalFiles: 1
- GatherStats: true
- MultiFileSpray: false
+ Select Operator
+ expressions:
+ expr: key
+ type: int
+ expr: value
+ type: string
+ outputColumnNames: _col0, _col1
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ directory: pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2011-07-11_02-58-23_160_5772419563904958236/-ext-10002
+ NumFilesPerFileSink: 1
+ Stats Publishing Key Prefix: pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2011-07-11_02-58-23_160_5772419563904958236/-ext-10000/
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ bucket_count -1
+ columns key,value
+ columns.types int:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1
+ name default.dest1
+ serialization.ddl struct dest1 { i32 key, string value}
+ serialization.format 1
+ serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ transient_lastDdlTime 1310378303
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.dest1
+ TotalFiles: 1
+ GatherStats: true
+ MultiFileSpray: false
Needs Tagging: false
Path -> Alias:
pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt [s]
Modified: hive/trunk/ql/src/test/results/clientpositive/sample8.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/sample8.q.out?rev=1146129&r1=1146128&r2=1146129&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/sample8.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/sample8.q.out Wed Jul 13 16:49:22 2011
@@ -33,49 +33,39 @@ STAGE PLANS:
alias: s
GatherStats: false
Filter Operator
- isSamplingPred: false
+ isSamplingPred: true
predicate:
expr: (((hash(key) & 2147483647) % 1) = 0)
type: boolean
- Filter Operator
- isSamplingPred: true
- predicate:
- expr: (((hash(key) & 2147483647) % 1) = 0)
- type: boolean
- Reduce Output Operator
- sort order:
- tag: 0
- value expressions:
- expr: key
- type: string
- expr: value
- type: string
- expr: ds
- type: string
- expr: hr
- type: string
+ Reduce Output Operator
+ sort order:
+ tag: 0
+ value expressions:
+ expr: key
+ type: string
+ expr: value
+ type: string
+ expr: ds
+ type: string
+ expr: hr
+ type: string
t
TableScan
alias: t
GatherStats: false
Filter Operator
- isSamplingPred: false
+ isSamplingPred: true
predicate:
expr: (((hash(key) & 2147483647) % 10) = 0)
type: boolean
- Filter Operator
- isSamplingPred: true
- predicate:
- expr: (((hash(key) & 2147483647) % 10) = 0)
- type: boolean
- Reduce Output Operator
- sort order:
- tag: 1
- value expressions:
- expr: key
- type: string
- expr: value
- type: string
+ Reduce Output Operator
+ sort order:
+ tag: 1
+ value expressions:
+ expr: key
+ type: string
+ expr: value
+ type: string
Needs Tagging: true
Path -> Alias:
pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [t, s]
Modified: hive/trunk/ql/src/test/results/clientpositive/sample9.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/sample9.q.out?rev=1146129&r1=1146128&r2=1146129&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/sample9.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/sample9.q.out Wed Jul 13 16:49:22 2011
@@ -22,52 +22,47 @@ STAGE PLANS:
alias: a
GatherStats: false
Filter Operator
- isSamplingPred: false
+ isSamplingPred: true
predicate:
expr: (((hash(key) & 2147483647) % 2) = 0)
type: boolean
- Filter Operator
- isSamplingPred: true
- predicate:
- expr: (((hash(key) & 2147483647) % 2) = 0)
- type: boolean
+ Select Operator
+ expressions:
+ expr: key
+ type: int
+ expr: value
+ type: string
+ outputColumnNames: _col0, _col1
Select Operator
expressions:
- expr: key
+ expr: _col0
type: int
- expr: value
+ expr: _col1
type: string
outputColumnNames: _col0, _col1
- Select Operator
- expressions:
- expr: _col0
- type: int
- expr: _col1
- type: string
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- GlobalTableId: 0
- directory: file:/tmp/sdong/hive_2011-04-19_19-55-15_093_8705959538820118982/-ext-10001
- NumFilesPerFileSink: 1
- Stats Publishing Key Prefix: file:/tmp/sdong/hive_2011-04-19_19-55-15_093_8705959538820118982/-ext-10001/
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- properties:
- columns _col0,_col1
- columns.types int:string
- serialization.format 1
- TotalFiles: 1
- GatherStats: false
- MultiFileSpray: false
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ directory: file:/tmp/amarsri/hive_2011-07-11_02-58-35_476_3346956740908751236/-ext-10001
+ NumFilesPerFileSink: 1
+ Stats Publishing Key Prefix: file:/tmp/amarsri/hive_2011-07-11_02-58-35_476_3346956740908751236/-ext-10001/
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ columns _col0,_col1
+ columns.types int:string
+ serialization.format 1
+ TotalFiles: 1
+ GatherStats: false
+ MultiFileSpray: false
Needs Tagging: false
Path -> Alias:
- pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt [s:a]
+ pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket [s:a]
Path -> Partition:
- pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt
+ pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket
Partition
- base file name: srcbucket0.txt
+ base file name: srcbucket
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
@@ -77,12 +72,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket
name default.srcbucket
serialization.ddl struct srcbucket { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303260542
+ transient_lastDdlTime 1310371847
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -94,12 +89,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcbucket
+ location pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket
name default.srcbucket
serialization.ddl struct srcbucket { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1303260542
+ transient_lastDdlTime 1310371847
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcbucket
name: default.srcbucket
@@ -113,12 +108,12 @@ PREHOOK: query: SELECT s.*
FROM (SELECT a.* FROM srcbucket TABLESAMPLE (BUCKET 1 OUT OF 2 on key) a) s
PREHOOK: type: QUERY
PREHOOK: Input: default@srcbucket
-PREHOOK: Output: file:/tmp/sdong/hive_2011-04-19_19-55-15_304_876052844698471163/-mr-10000
+PREHOOK: Output: file:/tmp/amarsri/hive_2011-07-11_02-58-35_521_5870581689660228083/-mr-10000
POSTHOOK: query: SELECT s.*
FROM (SELECT a.* FROM srcbucket TABLESAMPLE (BUCKET 1 OUT OF 2 on key) a) s
POSTHOOK: type: QUERY
POSTHOOK: Input: default@srcbucket
-POSTHOOK: Output: file:/tmp/sdong/hive_2011-04-19_19-55-15_304_876052844698471163/-mr-10000
+POSTHOOK: Output: file:/tmp/amarsri/hive_2011-07-11_02-58-35_521_5870581689660228083/-mr-10000
474 val_475
62 val_63
468 val_469