You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by nz...@apache.org on 2010/07/23 01:57:42 UTC
svn commit: r966909 [5/27] - in /hadoop/hive/trunk: ./ contrib/
data/warehouse/src/ hbase-handler/ hwi/ jdbc/ odbc/ ql/
ql/src/test/org/apache/hadoop/hive/ql/ ql/src/test/queries/clientnegative/
ql/src/test/queries/clientpositive/ ql/src/test/results/c...
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/bucket2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/bucket2.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/bucket2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/bucket2.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table bucket2_1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table bucket2_1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE bucket2_1(key int, value string) CLUSTERED BY (key) INTO 2 BUCKETS
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE bucket2_1(key int, value string) CLUSTERED BY (key) INTO 2 BUCKETS
@@ -49,9 +45,9 @@ STAGE PLANS:
type: string
Needs Tagging: false
Path -> Alias:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/test/data/warehouse/src [src]
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/src [src]
Path -> Partition:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/test/data/warehouse/src
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/src
Partition
base file name: src
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -62,12 +58,12 @@ STAGE PLANS:
columns.types string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/test/data/warehouse/src
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/src
name src
serialization.ddl struct src { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270515770
+ transient_lastDdlTime 1279735685
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -78,12 +74,12 @@ STAGE PLANS:
columns.types string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/test/data/warehouse/src
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/src
name src
serialization.ddl struct src { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270515770
+ transient_lastDdlTime 1279735685
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: src
name: src
@@ -99,7 +95,7 @@ STAGE PLANS:
File Output Operator
compressed: false
GlobalTableId: 1
- directory: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-02-51_261_8059892845101746655/10000
+ directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-09-11_789_5710176838021451402/10000
NumFilesPerFileSink: 2
table:
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -111,12 +107,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/test/data/warehouse/bucket2_1
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/bucket2_1
name bucket2_1
serialization.ddl struct bucket2_1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270515771
+ transient_lastDdlTime 1279735751
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucket2_1
TotalFiles: 2
@@ -126,7 +122,7 @@ STAGE PLANS:
Move Operator
tables:
replace: true
- source: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-02-51_261_8059892845101746655/10000
+ source: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-09-11_789_5710176838021451402/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -137,15 +133,15 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/test/data/warehouse/bucket2_1
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/bucket2_1
name bucket2_1
serialization.ddl struct bucket2_1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270515771
+ transient_lastDdlTime 1279735751
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucket2_1
- tmp directory: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-02-51_261_8059892845101746655/10001
+ tmp directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-09-11_789_5710176838021451402/10001
PREHOOK: query: insert overwrite table bucket2_1
@@ -225,11 +221,11 @@ STAGE PLANS:
PREHOOK: query: select * from bucket2_1 tablesample (bucket 1 out of 2) s order by key
PREHOOK: type: QUERY
PREHOOK: Input: default@bucket2_1
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-02-58_513_3181118859318846401/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-14_833_8954758164660929050/10000
POSTHOOK: query: select * from bucket2_1 tablesample (bucket 1 out of 2) s order by key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@bucket2_1
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-02-58_513_3181118859318846401/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-14_833_8954758164660929050/10000
POSTHOOK: Lineage: bucket2_1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: bucket2_1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
0 val_0
@@ -479,10 +475,3 @@ POSTHOOK: Lineage: bucket2_1.value SIMPL
498 val_498
498 val_498
498 val_498
-PREHOOK: query: drop table bucket2_1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table bucket2_1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@bucket2_1
-POSTHOOK: Lineage: bucket2_1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: bucket2_1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/bucket3.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/bucket3.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/bucket3.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/bucket3.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table bucket3_1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table bucket3_1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE bucket3_1(key int, value string) partitioned by (ds string) CLUSTERED BY (key) INTO 2 BUCKETS
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE bucket3_1(key int, value string) partitioned by (ds string) CLUSTERED BY (key) INTO 2 BUCKETS
@@ -49,9 +45,9 @@ STAGE PLANS:
type: string
Needs Tagging: false
Path -> Alias:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/test/data/warehouse/src [src]
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/src [src]
Path -> Partition:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/test/data/warehouse/src
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/src
Partition
base file name: src
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -62,12 +58,12 @@ STAGE PLANS:
columns.types string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/test/data/warehouse/src
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/src
name src
serialization.ddl struct src { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270515793
+ transient_lastDdlTime 1279735685
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -78,12 +74,12 @@ STAGE PLANS:
columns.types string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/test/data/warehouse/src
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/src
name src
serialization.ddl struct src { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270515793
+ transient_lastDdlTime 1279735685
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: src
name: src
@@ -99,7 +95,7 @@ STAGE PLANS:
File Output Operator
compressed: false
GlobalTableId: 1
- directory: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-03-14_419_5303321979362543224/10000
+ directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-09-18_069_2499676602955827966/10000
NumFilesPerFileSink: 2
table:
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -111,13 +107,13 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/test/data/warehouse/bucket3_1
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/bucket3_1
name bucket3_1
partition_columns ds
serialization.ddl struct bucket3_1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270515794
+ transient_lastDdlTime 1279735757
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucket3_1
TotalFiles: 2
@@ -129,7 +125,7 @@ STAGE PLANS:
partition:
ds 1
replace: true
- source: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-03-14_419_5303321979362543224/10000
+ source: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-09-18_069_2499676602955827966/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -140,16 +136,16 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/test/data/warehouse/bucket3_1
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/bucket3_1
name bucket3_1
partition_columns ds
serialization.ddl struct bucket3_1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270515794
+ transient_lastDdlTime 1279735757
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucket3_1
- tmp directory: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-03-14_419_5303321979362543224/10001
+ tmp directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-09-18_069_2499676602955827966/10001
PREHOOK: query: insert overwrite table bucket3_1 partition (ds='1')
@@ -253,11 +249,11 @@ STAGE PLANS:
PREHOOK: query: select * from bucket3_1 tablesample (bucket 1 out of 2) s where ds = '1' order by key
PREHOOK: type: QUERY
PREHOOK: Input: default@bucket3_1@ds=1
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-03-24_389_3958955156388135589/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-23_782_5794755668640533078/10000
POSTHOOK: query: select * from bucket3_1 tablesample (bucket 1 out of 2) s where ds = '1' order by key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@bucket3_1@ds=1
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-03-24_389_3958955156388135589/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-23_782_5794755668640533078/10000
POSTHOOK: Lineage: bucket3_1 PARTITION(ds=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: bucket3_1 PARTITION(ds=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: bucket3_1 PARTITION(ds=2).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
@@ -509,12 +505,3 @@ POSTHOOK: Lineage: bucket3_1 PARTITION(d
498 val_498 1
498 val_498 1
498 val_498 1
-PREHOOK: query: drop table bucket3_1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table bucket3_1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@bucket3_1
-POSTHOOK: Lineage: bucket3_1 PARTITION(ds=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: bucket3_1 PARTITION(ds=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: bucket3_1 PARTITION(ds=2).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: bucket3_1 PARTITION(ds=2).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/bucket4.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/bucket4.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/bucket4.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/bucket4.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table bucket4_1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table bucket4_1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE bucket4_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE bucket4_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS
@@ -52,9 +48,9 @@ STAGE PLANS:
type: string
Needs Tagging: false
Path -> Alias:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/src [src]
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/src [src]
Path -> Partition:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/src
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/src
Partition
base file name: src
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -65,12 +61,12 @@ STAGE PLANS:
columns.types string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/src
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/src
name src
serialization.ddl struct src { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270515769
+ transient_lastDdlTime 1279735685
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -81,12 +77,12 @@ STAGE PLANS:
columns.types string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/src
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/src
name src
serialization.ddl struct src { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270515769
+ transient_lastDdlTime 1279735685
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: src
name: src
@@ -102,7 +98,7 @@ STAGE PLANS:
File Output Operator
compressed: false
GlobalTableId: 1
- directory: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-02-50_258_5147459717740143983/10000
+ directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-09-26_887_4786734363166528929/10000
NumFilesPerFileSink: 2
table:
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -115,12 +111,12 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/bucket4_1
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/bucket4_1
name bucket4_1
serialization.ddl struct bucket4_1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270515770
+ transient_lastDdlTime 1279735766
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucket4_1
TotalFiles: 2
@@ -130,7 +126,7 @@ STAGE PLANS:
Move Operator
tables:
replace: true
- source: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-02-50_258_5147459717740143983/10000
+ source: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-09-26_887_4786734363166528929/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -142,15 +138,15 @@ STAGE PLANS:
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/bucket4_1
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/bucket4_1
name bucket4_1
serialization.ddl struct bucket4_1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270515770
+ transient_lastDdlTime 1279735766
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucket4_1
- tmp directory: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-02-50_258_5147459717740143983/10001
+ tmp directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-09-26_887_4786734363166528929/10001
PREHOOK: query: insert overwrite table bucket4_1
@@ -217,11 +213,11 @@ STAGE PLANS:
PREHOOK: query: select * from bucket4_1 tablesample (bucket 1 out of 2) s
PREHOOK: type: QUERY
PREHOOK: Input: default@bucket4_1
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-02-55_285_753310371043040764/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-29_642_3068318549775797820/10000
POSTHOOK: query: select * from bucket4_1 tablesample (bucket 1 out of 2) s
POSTHOOK: type: QUERY
POSTHOOK: Input: default@bucket4_1
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-02-55_285_753310371043040764/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-29_642_3068318549775797820/10000
POSTHOOK: Lineage: bucket4_1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: bucket4_1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
0 val_0
@@ -471,10 +467,3 @@ POSTHOOK: Lineage: bucket4_1.value SIMPL
498 val_498
498 val_498
498 val_498
-PREHOOK: query: drop table bucket4_1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table bucket4_1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@bucket4_1
-POSTHOOK: Lineage: bucket4_1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: bucket4_1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/bucket_groupby.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/bucket_groupby.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/bucket_groupby.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/bucket_groupby.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table clustergroupby
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table clustergroupby
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table clustergroupby(key string, value string) partitioned by(ds string)
PREHOOK: type: CREATETABLE
POSTHOOK: query: create table clustergroupby(key string, value string) partitioned by(ds string)
@@ -15,7 +11,7 @@ key string
value string
ds string
-Detailed Table Information Table(tableName:clustergroupby, dbName:default, owner:athusoo, createTime:1270515787, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/test/data/warehouse/clustergroupby, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null)], parameters:{transient_lastDdlTime=1270515787}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)
+Detailed Table Information Table(tableName:clustergroupby, dbName:default, owner:jssarma, createTime:1279735772, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/clustergroupby, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null)], parameters:{transient_lastDdlTime=1279735772}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)
PREHOOK: query: insert overwrite table clustergroupby partition (ds='100') select key, value from src sort by key
PREHOOK: type: QUERY
PREHOOK: Input: default@src
@@ -115,11 +111,11 @@ STAGE PLANS:
PREHOOK: query: select key, count(1) from clustergroupby where ds='100' group by key limit 10
PREHOOK: type: QUERY
PREHOOK: Input: default@clustergroupby@ds=100
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-03-13_780_4293831568302982785/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-35_534_8185562718357069120/10000
POSTHOOK: query: select key, count(1) from clustergroupby where ds='100' group by key limit 10
POSTHOOK: type: QUERY
POSTHOOK: Input: default@clustergroupby@ds=100
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-03-13_780_4293831568302982785/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-35_534_8185562718357069120/10000
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
0 3
@@ -150,7 +146,7 @@ key string
value string
ds string
-Detailed Table Information Table(tableName:clustergroupby, dbName:default, owner:athusoo, createTime:1270515787, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/test/data/warehouse/clustergroupby, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[key], sortCols:[], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null)], parameters:{last_modified_by=athusoo,last_modified_time=1270515798,transient_lastDdlTime=1270515798}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)
+Detailed Table Information Table(tableName:clustergroupby, dbName:default, owner:jssarma, createTime:1279735772, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/clustergroupby, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[key], sortCols:[], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null)], parameters:{last_modified_by=jssarma, last_modified_time=1279735778, transient_lastDdlTime=1279735778}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)
PREHOOK: query: insert overwrite table clustergroupby partition (ds='101') select key, value from src distribute by key
PREHOOK: type: QUERY
PREHOOK: Input: default@src
@@ -256,11 +252,11 @@ STAGE PLANS:
PREHOOK: query: select key, count(1) from clustergroupby where ds='101' group by key limit 10
PREHOOK: type: QUERY
PREHOOK: Input: default@clustergroupby@ds=101
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-03-24_380_6328201988353238872/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-41_016_7689771879965686688/10000
POSTHOOK: query: select key, count(1) from clustergroupby where ds='101' group by key limit 10
POSTHOOK: type: QUERY
POSTHOOK: Input: default@clustergroupby@ds=101
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-03-24_380_6328201988353238872/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-41_016_7689771879965686688/10000
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=101).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
@@ -368,11 +364,11 @@ STAGE PLANS:
PREHOOK: query: select length(key), count(1) from clustergroupby where ds='101' group by length(key) limit 10
PREHOOK: type: QUERY
PREHOOK: Input: default@clustergroupby@ds=101
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-03-30_930_3497817425124433255/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-43_849_707604862542797827/10000
POSTHOOK: query: select length(key), count(1) from clustergroupby where ds='101' group by length(key) limit 10
POSTHOOK: type: QUERY
POSTHOOK: Input: default@clustergroupby@ds=101
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-03-30_930_3497817425124433255/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-43_849_707604862542797827/10000
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=101).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
@@ -471,11 +467,11 @@ STAGE PLANS:
PREHOOK: query: select abs(length(key)), count(1) from clustergroupby where ds='101' group by abs(length(key)) limit 10
PREHOOK: type: QUERY
PREHOOK: Input: default@clustergroupby@ds=101
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-03-41_422_5290311054247228986/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-46_664_3688496316777330614/10000
POSTHOOK: query: select abs(length(key)), count(1) from clustergroupby where ds='101' group by abs(length(key)) limit 10
POSTHOOK: type: QUERY
POSTHOOK: Input: default@clustergroupby@ds=101
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-03-41_422_5290311054247228986/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-46_664_3688496316777330614/10000
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=101).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
@@ -584,11 +580,11 @@ STAGE PLANS:
PREHOOK: query: select key, count(1) from clustergroupby where ds='101' group by key,3 limit 10
PREHOOK: type: QUERY
PREHOOK: Input: default@clustergroupby@ds=101
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-03-47_388_6623885853323229647/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-49_481_9074398325986135965/10000
POSTHOOK: query: select key, count(1) from clustergroupby where ds='101' group by key,3 limit 10
POSTHOOK: type: QUERY
POSTHOOK: Input: default@clustergroupby@ds=101
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-03-47_388_6623885853323229647/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-49_481_9074398325986135965/10000
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=101).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
@@ -701,11 +697,11 @@ STAGE PLANS:
PREHOOK: query: select key, count(1) from (select value as key, key as value from clustergroupby where ds='101')subq group by key limit 10
PREHOOK: type: QUERY
PREHOOK: Input: default@clustergroupby@ds=101
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-03-55_205_578266570097178049/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-52_312_6910555695381377185/10000
POSTHOOK: query: select key, count(1) from (select value as key, key as value from clustergroupby where ds='101')subq group by key limit 10
POSTHOOK: type: QUERY
POSTHOOK: Input: default@clustergroupby@ds=101
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-03-55_205_578266570097178049/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-52_312_6910555695381377185/10000
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=101).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
@@ -803,12 +799,12 @@ PREHOOK: query: select key, count(1) fro
PREHOOK: type: QUERY
PREHOOK: Input: default@clustergroupby@ds=100
PREHOOK: Input: default@clustergroupby@ds=101
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-04-08_207_8200567238299494473/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-55_221_4522384383217894505/10000
POSTHOOK: query: select key, count(1) from clustergroupby group by key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@clustergroupby@ds=100
POSTHOOK: Input: default@clustergroupby@ds=101
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-04-08_207_8200567238299494473/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-55_221_4522384383217894505/10000
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=101).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
@@ -1233,7 +1229,7 @@ key string
value string
ds string
-Detailed Table Information Table(tableName:clustergroupby, dbName:default, owner:athusoo, createTime:1270515787, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/test/data/warehouse/clustergroupby, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[value], sortCols:[Order(col:key, order:1), Order(col:value, order:1)], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null)], parameters:{last_modified_by=athusoo,last_modified_time=1270515867,transient_lastDdlTime=1270515867}, viewOriginalText:null, viewExpa
ndedText:null, tableType:MANAGED_TABLE)
+Detailed Table Information Table(tableName:clustergroupby, dbName:default, owner:jssarma, createTime:1279735772, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/clustergroupby, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[value], sortCols:[Order(col:key, order:1), Order(col:value, order:1)], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null)], parameters:{last_modified_by=jssarma, last_modified_time=1279735798, transient_lastDdlTime=1279735798}, viewOriginalText:null, viewExpandedText:
null, tableType:MANAGED_TABLE)
PREHOOK: query: insert overwrite table clustergroupby partition (ds='102') select key, value from src distribute by value sort by key, value
PREHOOK: type: QUERY
PREHOOK: Input: default@src
@@ -1341,11 +1337,11 @@ STAGE PLANS:
PREHOOK: query: select key, count(1) from clustergroupby where ds='102' group by key limit 10
PREHOOK: type: QUERY
PREHOOK: Input: default@clustergroupby@ds=102
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-04-33_613_3501792825410171493/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-10-01_215_2690537963790427608/10000
POSTHOOK: query: select key, count(1) from clustergroupby where ds='102' group by key limit 10
POSTHOOK: type: QUERY
POSTHOOK: Input: default@clustergroupby@ds=102
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-04-33_613_3501792825410171493/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-10-01_215_2690537963790427608/10000
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=101).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
@@ -1455,11 +1451,11 @@ STAGE PLANS:
PREHOOK: query: select value, count(1) from clustergroupby where ds='102' group by value limit 10
PREHOOK: type: QUERY
PREHOOK: Input: default@clustergroupby@ds=102
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-04-38_518_2690806199946056325/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-10-04_135_3923631383708910626/10000
POSTHOOK: query: select value, count(1) from clustergroupby where ds='102' group by value limit 10
POSTHOOK: type: QUERY
POSTHOOK: Input: default@clustergroupby@ds=102
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-04-38_518_2690806199946056325/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-10-04_135_3923631383708910626/10000
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=101).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
@@ -1579,11 +1575,11 @@ STAGE PLANS:
PREHOOK: query: select key, count(1) from clustergroupby where ds='102' group by key, value limit 10
PREHOOK: type: QUERY
PREHOOK: Input: default@clustergroupby@ds=102
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-04-43_716_5095162971444718926/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-10-07_055_4285817181416116197/10000
POSTHOOK: query: select key, count(1) from clustergroupby where ds='102' group by key, value limit 10
POSTHOOK: type: QUERY
POSTHOOK: Input: default@clustergroupby@ds=102
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-04-43_716_5095162971444718926/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-10-07_055_4285817181416116197/10000
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=101).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
@@ -1626,7 +1622,7 @@ key string
value string
ds string
-Detailed Table Information Table(tableName:clustergroupby, dbName:default, owner:athusoo, createTime:1270515787, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/test/data/warehouse/clustergroupby, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[value, key], sortCols:[Order(col:key, order:1)], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null)], parameters:{last_modified_by=athusoo,last_modified_time=1270515888,transient_lastDdlTime=1270515888}, viewOriginalText:null, viewExpandedText:null, tableTy
pe:MANAGED_TABLE)
+Detailed Table Information Table(tableName:clustergroupby, dbName:default, owner:jssarma, createTime:1279735772, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/clustergroupby, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[value, key], sortCols:[Order(col:key, order:1)], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null)], parameters:{last_modified_by=jssarma, last_modified_time=1279735809, transient_lastDdlTime=1279735809}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGE
D_TABLE)
PREHOOK: query: insert overwrite table clustergroupby partition (ds='103') select key, value from src distribute by value, key sort by key
PREHOOK: type: QUERY
PREHOOK: Input: default@src
@@ -1738,11 +1734,11 @@ STAGE PLANS:
PREHOOK: query: select key, count(1) from clustergroupby where ds='103' group by key limit 10
PREHOOK: type: QUERY
PREHOOK: Input: default@clustergroupby@ds=103
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-04-53_057_86001272118042631/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-10-12_709_7466980661091036250/10000
POSTHOOK: query: select key, count(1) from clustergroupby where ds='103' group by key limit 10
POSTHOOK: type: QUERY
POSTHOOK: Input: default@clustergroupby@ds=103
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-04-53_057_86001272118042631/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-10-12_709_7466980661091036250/10000
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=101).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
@@ -1866,11 +1862,11 @@ STAGE PLANS:
PREHOOK: query: select key, count(1) from clustergroupby where ds='103' group by value, key limit 10
PREHOOK: type: QUERY
PREHOOK: Input: default@clustergroupby@ds=103
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-04-57_724_2028785942026261985/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-10-15_551_6887415689280301293/10000
POSTHOOK: query: select key, count(1) from clustergroupby where ds='103' group by value, key limit 10
POSTHOOK: type: QUERY
POSTHOOK: Input: default@clustergroupby@ds=103
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-04-57_724_2028785942026261985/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-10-15_551_6887415689280301293/10000
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: clustergroupby PARTITION(ds=101).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
@@ -1889,16 +1885,3 @@ POSTHOOK: Lineage: clustergroupby PARTIT
111 1
113 2
114 1
-PREHOOK: query: drop table clustergroupby
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table clustergroupby
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@clustergroupby
-POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: clustergroupby PARTITION(ds=101).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: clustergroupby PARTITION(ds=101).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: clustergroupby PARTITION(ds=102).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: clustergroupby PARTITION(ds=102).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: clustergroupby PARTITION(ds=103).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: clustergroupby PARTITION(ds=103).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/bucketizedhiveinputformat.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/bucketizedhiveinputformat.q.out?rev=966909&r1=966908&r2=966909&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/bucketizedhiveinputformat.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/bucketizedhiveinputformat.q.out Thu Jul 22 23:57:29 2010
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE T1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE T1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE T1(name STRING) STORED AS TEXTFILE
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE T1(name STRING) STORED AS TEXTFILE
@@ -12,10 +8,6 @@ PREHOOK: type: LOAD
POSTHOOK: query: LOAD DATA LOCAL INPATH '../data/files/kv1.txt' INTO TABLE T1
POSTHOOK: type: LOAD
POSTHOOK: Output: default@t1
-PREHOOK: query: DROP TABLE T2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE T2
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE T2(name STRING) STORED AS SEQUENCEFILE
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE T2(name STRING) STORED AS SEQUENCEFILE
@@ -163,7 +155,7 @@ STAGE PLANS:
Stage: Stage-3
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-03-34_477_4466762848424481616/10003
+ file:/tmp/jssarma/hive_2010-07-21_11-10-19_047_380113812021070738/10003
Reduce Output Operator
sort order:
tag: -1
@@ -271,18 +263,13 @@ STAGE PLANS:
PREHOOK: query: SELECT COUNT(1) FROM T2
PREHOOK: type: QUERY
PREHOOK: Input: default@t2
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-06-18_958_8326164509200163533/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-11-40_372_2216491831322314327/10000
POSTHOOK: query: SELECT COUNT(1) FROM T2
POSTHOOK: type: QUERY
POSTHOOK: Input: default@t2
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-06-18_958_8326164509200163533/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-11-40_372_2216491831322314327/10000
POSTHOOK: Lineage: t2.name SIMPLE [(t1)t1.FieldSchema(name:name, type:string, comment:null), ]
5000000
-PREHOOK: query: DROP TABLE T3
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE T3
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Lineage: t2.name SIMPLE [(t1)t1.FieldSchema(name:name, type:string, comment:null), ]
PREHOOK: query: CREATE TABLE T3(name STRING) STORED AS TEXTFILE
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE T3(name STRING) STORED AS TEXTFILE
@@ -360,28 +347,10 @@ STAGE PLANS:
PREHOOK: query: SELECT COUNT(1) FROM T3
PREHOOK: type: QUERY
PREHOOK: Input: default@t3
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-06-45_448_8066146192994060790/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-11-52_373_3173950263104621965/10000
POSTHOOK: query: SELECT COUNT(1) FROM T3
POSTHOOK: type: QUERY
POSTHOOK: Input: default@t3
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-06-45_448_8066146192994060790/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-11-52_373_3173950263104621965/10000
POSTHOOK: Lineage: t2.name SIMPLE [(t1)t1.FieldSchema(name:name, type:string, comment:null), ]
1000
-PREHOOK: query: DROP TABLE T1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE T1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@t1
-POSTHOOK: Lineage: t2.name SIMPLE [(t1)t1.FieldSchema(name:name, type:string, comment:null), ]
-PREHOOK: query: DROP TABLE T2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE T2
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@t2
-POSTHOOK: Lineage: t2.name SIMPLE [(t1)t1.FieldSchema(name:name, type:string, comment:null), ]
-PREHOOK: query: DROP TABLE T3
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE T3
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@t3
-POSTHOOK: Lineage: t2.name SIMPLE [(t1)t1.FieldSchema(name:name, type:string, comment:null), ]