You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by nz...@apache.org on 2010/02/14 03:31:44 UTC
svn commit: r909965 [1/15] - in /hadoop/hive/trunk: ./
contrib/src/test/results/clientnegative/
contrib/src/test/results/clientpositive/
ql/src/java/org/apache/hadoop/hive/ql/exec/
ql/src/test/results/clientpositive/
Author: nzhang
Date: Sun Feb 14 02:31:40 2010
New Revision: 909965
URL: http://svn.apache.org/viewvc?rev=909965&view=rev
Log:
HIVE-1167. Use TreeMap instead of Property to make explain extended deterministic (Zheng Shao via Ning Zhang)
Modified:
hadoop/hive/trunk/CHANGES.txt
hadoop/hive/trunk/contrib/src/test/results/clientnegative/serde_regex.q.out
hadoop/hive/trunk/contrib/src/test/results/clientpositive/serde_regex.q.out
hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExplainTask.java
hadoop/hive/trunk/ql/src/test/results/clientpositive/binary_output_format.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/cluster.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/ctas.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby_map_ppr.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby_ppr.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/input23.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/input39.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/input42.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/input_part1.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/input_part2.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/input_part7.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/input_part9.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join0.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join1.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join11.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join12.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join13.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join14.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join15.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join17.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join19.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join2.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join20.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join21.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join22.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join23.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join26.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join3.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join32.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join33.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join34.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join35.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join40.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join9.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join_hive_626.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join_map_ppr.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join_rc.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder2.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder3.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/join_thrift.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/louter_join_ppr.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/multi_insert.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/no_hooks.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/outer_join_ppr.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_clusterby.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_gby_join.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_join.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_join2.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_join3.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_multi_insert.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join1.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join2.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join3.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join4.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_random.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_udf_case.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner1.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner2.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner3.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/regex_col.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/regexp_extract.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/router_join_ppr.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/sample1.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/sample2.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/sample4.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/sample5.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/sample6.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/sample8.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/sample9.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/semijoin.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/skewjoin.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/transform_ppr1.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/transform_ppr2.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case_column_pruning.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_explode.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/union16.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/union20.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/union21.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/union22.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/union23.q.out
hadoop/hive/trunk/ql/src/test/results/clientpositive/union_ppr.q.out
Modified: hadoop/hive/trunk/CHANGES.txt
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/CHANGES.txt?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/CHANGES.txt (original)
+++ hadoop/hive/trunk/CHANGES.txt Sun Feb 14 02:31:40 2010
@@ -166,6 +166,10 @@
HIVE-1164. Fix drop_partition_by_name. (Paul Yang via zshao)
+ HIVE-1167. Use TreeMap instead of Property to make explain extended
+ deterministic
+ (Zheng Shao via Ning Zhang)
+
Release 0.5.0 - Unreleased
INCOMPATIBLE CHANGES
Modified: hadoop/hive/trunk/contrib/src/test/results/clientnegative/serde_regex.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/contrib/src/test/results/clientnegative/serde_regex.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/contrib/src/test/results/clientnegative/serde_regex.q.out (original)
+++ hadoop/hive/trunk/contrib/src/test/results/clientnegative/serde_regex.q.out Sun Feb 14 02:31:40 2010
@@ -52,8 +52,8 @@
if not exists: false
input format: org.apache.hadoop.mapred.TextInputFormat
serde properties:
- output.format.string %1$s %2$s %3$s %4$s %5$s %6$s %7$s %8$s %9$s
input.regex ([^ ]*) ([^ ]*) ([^ ]*) (-|\[[^\]]*\]) ([^ "]*|"[^"]*") (-|[0-9]*) (-|[0-9]*)(?: ([^ "]*|"[^"]*") ([^ "]*|"[^"]*"))?
+ output.format.string %1$s %2$s %3$s %4$s %5$s %6$s %7$s %8$s %9$s
# buckets: -1
output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
serde name: org.apache.hadoop.hive.contrib.serde2.RegexSerDe
Modified: hadoop/hive/trunk/contrib/src/test/results/clientpositive/serde_regex.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/contrib/src/test/results/clientpositive/serde_regex.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/contrib/src/test/results/clientpositive/serde_regex.q.out (original)
+++ hadoop/hive/trunk/contrib/src/test/results/clientpositive/serde_regex.q.out Sun Feb 14 02:31:40 2010
@@ -52,8 +52,8 @@
if not exists: false
input format: org.apache.hadoop.mapred.TextInputFormat
serde properties:
- output.format.string %1$s %2$s %3$s %4$s %5$s %6$s %7$s %8$s %9$s
input.regex ([^ ]*) ([^ ]*) ([^ ]*) (-|\[[^\]]*\]) ([^ "]*|"[^"]*") (-|[0-9]*) (-|[0-9]*)(?: ([^ "]*|"[^"]*") ([^ "]*|"[^"]*"))?
+ output.format.string %1$s %2$s %3$s %4$s %5$s %6$s %7$s %8$s %9$s
# buckets: -1
output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
serde name: org.apache.hadoop.hive.contrib.serde2.RegexSerDe
@@ -109,11 +109,11 @@
PREHOOK: query: SELECT * FROM serde_regex ORDER BY time
PREHOOK: type: QUERY
PREHOOK: Input: default@serde_regex
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/569581247/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_3/build/contrib/scratchdir/hive_2010-02-12_22-33-19_616_3967776384774299534/10000
POSTHOOK: query: SELECT * FROM serde_regex ORDER BY time
POSTHOOK: type: QUERY
POSTHOOK: Input: default@serde_regex
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/569581247/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_3/build/contrib/scratchdir/hive_2010-02-12_22-33-19_616_3967776384774299534/10000
127.0.0.1 - frank [10/Oct/2000:13:55:36 -0700] "GET /apache_pb.gif HTTP/1.0" 200 2326 NULL NULL
127.0.0.1 - - [26/May/2009:00:00:00 +0000] "GET /someurl/?track=Blabla(Main) HTTP/1.1" 200 5864 - "Mozilla/5.0 (Windows; U; Windows NT 6.0; en-US) AppleWebKit/525.19 (KHTML, like Gecko) Chrome/1.0.154.65 Safari/525.19"
PREHOOK: query: DROP TABLE serde_regex
Modified: hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExplainTask.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExplainTask.java?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExplainTask.java (original)
+++ hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExplainTask.java Sun Feb 14 02:31:40 2010
@@ -29,6 +29,7 @@
import java.util.List;
import java.util.Map;
import java.util.Set;
+import java.util.TreeMap;
import java.util.Map.Entry;
import org.apache.hadoop.fs.Path;
@@ -88,7 +89,9 @@
boolean extended, int indent) throws Exception {
boolean first_el = true;
- for (Entry<?, ?> ent : mp.entrySet()) {
+ TreeMap<Object, Object> tree = new TreeMap<Object, Object>();
+ tree.putAll(mp);
+ for (Entry<?, ?> ent : tree.entrySet()) {
if (first_el) {
out.println(header);
}
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/binary_output_format.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/binary_output_format.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/binary_output_format.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/binary_output_format.q.out Sun Feb 14 02:31:40 2010
@@ -83,71 +83,71 @@
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- field.delim 9
columns _col0
- serialization.use.json.object true
- serialization.format 9
columns.types string
+ field.delim 9
+ serialization.format 9
serialization.last.column.takes.rest true
+ serialization.use.json.object true
File Output Operator
compressed: false
GlobalTableId: 1
- directory: file:/data/users/zshao/svnhive_hadoop/trunk/VENDOR.hive/trunk/build/ql/scratchdir/1286801631/10002
+ directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-15-58_271_3464965796195557656/10002
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
properties:
- name dest1
+ bucket_count -1
+ columns mydata
columns.types string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/dest1
+ name dest1
serialization.ddl struct dest1 { string mydata}
- columns mydata
serialization.format 1
serialization.last.column.takes.rest true
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
- location file:/data/users/zshao/svnhive_hadoop/trunk/VENDOR.hive/trunk/build/ql/test/data/warehouse/dest1
- transient_lastDdlTime 1264211336
+ transient_lastDdlTime 1266041758
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest1
Needs Tagging: false
Path -> Alias:
- file:/data/users/zshao/svnhive_hadoop/trunk/VENDOR.hive/trunk/build/ql/test/data/warehouse/src [src]
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/src [src]
Path -> Partition:
- file:/data/users/zshao/svnhive_hadoop/trunk/VENDOR.hive/trunk/build/ql/test/data/warehouse/src
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/src
Partition
base file name: src
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name src
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/src
+ name src
serialization.ddl struct src { string key, string value}
serialization.format 1
- columns key,value
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/zshao/svnhive_hadoop/trunk/VENDOR.hive/trunk/build/ql/test/data/warehouse/src
- transient_lastDdlTime 1264211335
+ transient_lastDdlTime 1266041757
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name src
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/src
+ name src
serialization.ddl struct src { string key, string value}
serialization.format 1
- columns key,value
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/zshao/svnhive_hadoop/trunk/VENDOR.hive/trunk/build/ql/test/data/warehouse/src
- transient_lastDdlTime 1264211335
+ transient_lastDdlTime 1266041757
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: src
name: src
@@ -159,38 +159,38 @@
Move Operator
files:
hdfs directory: true
- source: file:/data/users/zshao/svnhive_hadoop/trunk/VENDOR.hive/trunk/build/ql/scratchdir/1286801631/10002
- destination: file:/data/users/zshao/svnhive_hadoop/trunk/VENDOR.hive/trunk/build/ql/scratchdir/1065869242/10000
+ source: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-15-58_271_3464965796195557656/10002
+ destination: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-15-58_271_3464965796195557656/10000
Stage: Stage-0
Move Operator
tables:
replace: true
- source: file:/data/users/zshao/svnhive_hadoop/trunk/VENDOR.hive/trunk/build/ql/scratchdir/1065869242/10000
+ source: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-15-58_271_3464965796195557656/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
properties:
- name dest1
+ bucket_count -1
+ columns mydata
columns.types string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/dest1
+ name dest1
serialization.ddl struct dest1 { string mydata}
- columns mydata
serialization.format 1
serialization.last.column.takes.rest true
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
- location file:/data/users/zshao/svnhive_hadoop/trunk/VENDOR.hive/trunk/build/ql/test/data/warehouse/dest1
- transient_lastDdlTime 1264211336
+ transient_lastDdlTime 1266041758
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest1
- tmp directory: file:/data/users/zshao/svnhive_hadoop/trunk/VENDOR.hive/trunk/build/ql/scratchdir/1065869242/10001
+ tmp directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-15-58_271_3464965796195557656/10001
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/zshao/svnhive_hadoop/trunk/VENDOR.hive/trunk/build/ql/scratchdir/1286801631/10002
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-15-58_271_3464965796195557656/10002
Reduce Output Operator
sort order:
Map-reduce partition columns:
@@ -202,43 +202,43 @@
type: string
Needs Tagging: false
Path -> Alias:
- file:/data/users/zshao/svnhive_hadoop/trunk/VENDOR.hive/trunk/build/ql/scratchdir/1286801631/10002 [file:/data/users/zshao/svnhive_hadoop/trunk/VENDOR.hive/trunk/build/ql/scratchdir/1286801631/10002]
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-15-58_271_3464965796195557656/10002 [file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-15-58_271_3464965796195557656/10002]
Path -> Partition:
- file:/data/users/zshao/svnhive_hadoop/trunk/VENDOR.hive/trunk/build/ql/scratchdir/1286801631/10002
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-15-58_271_3464965796195557656/10002
Partition
base file name: 10002
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
properties:
- name dest1
+ bucket_count -1
+ columns mydata
columns.types string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/dest1
+ name dest1
serialization.ddl struct dest1 { string mydata}
- columns mydata
serialization.format 1
serialization.last.column.takes.rest true
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
- location file:/data/users/zshao/svnhive_hadoop/trunk/VENDOR.hive/trunk/build/ql/test/data/warehouse/dest1
- transient_lastDdlTime 1264211336
+ transient_lastDdlTime 1266041758
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
properties:
- name dest1
+ bucket_count -1
+ columns mydata
columns.types string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/dest1
+ name dest1
serialization.ddl struct dest1 { string mydata}
- columns mydata
serialization.format 1
serialization.last.column.takes.rest true
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
- location file:/data/users/zshao/svnhive_hadoop/trunk/VENDOR.hive/trunk/build/ql/test/data/warehouse/dest1
- transient_lastDdlTime 1264211336
+ transient_lastDdlTime 1266041758
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest1
name: dest1
@@ -247,23 +247,23 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/zshao/svnhive_hadoop/trunk/VENDOR.hive/trunk/build/ql/scratchdir/1065869242/10000
+ directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-15-58_271_3464965796195557656/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
properties:
- name dest1
+ bucket_count -1
+ columns mydata
columns.types string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/dest1
+ name dest1
serialization.ddl struct dest1 { string mydata}
serialization.format 1
- columns mydata
serialization.last.column.takes.rest true
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- location file:/data/users/zshao/svnhive_hadoop/trunk/VENDOR.hive/trunk/build/ql/test/data/warehouse/dest1
- file.outputformat org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat
- transient_lastDdlTime 1264211336
+ transient_lastDdlTime 1266041758
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest1
@@ -300,12 +300,12 @@
SELECT * FROM dest1
PREHOOK: type: QUERY
PREHOOK: Input: default@dest1
-PREHOOK: Output: file:/data/users/zshao/svnhive_hadoop/trunk/VENDOR.hive/trunk/build/ql/scratchdir/1904733215/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-16-03_636_3562459403060191123/10000
POSTHOOK: query: -- Test the result
SELECT * FROM dest1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest1
-POSTHOOK: Output: file:/data/users/zshao/svnhive_hadoop/trunk/VENDOR.hive/trunk/build/ql/scratchdir/1904733215/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-16-03_636_3562459403060191123/10000
238 val_238
86 val_86
311 val_311
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/cluster.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/cluster.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/cluster.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/cluster.q.out Sun Feb 14 02:31:40 2010
@@ -64,11 +64,11 @@
PREHOOK: query: SELECT * FROM SRC x where x.key = 10 CLUSTER BY x.key
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/666819916/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-30_987_946002006407811313/10000
POSTHOOK: query: SELECT * FROM SRC x where x.key = 10 CLUSTER BY x.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/666819916/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-30_987_946002006407811313/10000
10 val_10
PREHOOK: query: EXPLAIN
SELECT * FROM SRC x where x.key = 20 CLUSTER BY key
@@ -136,11 +136,11 @@
PREHOOK: query: SELECT * FROM SRC x where x.key = 20 CLUSTER BY key
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/437686258/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-36_798_7722926056682905425/10000
POSTHOOK: query: SELECT * FROM SRC x where x.key = 20 CLUSTER BY key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/437686258/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-36_798_7722926056682905425/10000
20 val_20
PREHOOK: query: EXPLAIN
SELECT x.* FROM SRC x where x.key = 20 CLUSTER BY key
@@ -208,11 +208,11 @@
PREHOOK: query: SELECT x.* FROM SRC x where x.key = 20 CLUSTER BY key
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/479850350/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-43_717_6012279607867421749/10000
POSTHOOK: query: SELECT x.* FROM SRC x where x.key = 20 CLUSTER BY key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/479850350/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-43_717_6012279607867421749/10000
20 val_20
PREHOOK: query: EXPLAIN
SELECT x.* FROM SRC x where x.key = 20 CLUSTER BY x.key
@@ -280,11 +280,11 @@
PREHOOK: query: SELECT x.* FROM SRC x where x.key = 20 CLUSTER BY x.key
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/261160281/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-49_882_6526844180884182893/10000
POSTHOOK: query: SELECT x.* FROM SRC x where x.key = 20 CLUSTER BY x.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/261160281/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-49_882_6526844180884182893/10000
20 val_20
PREHOOK: query: EXPLAIN
SELECT x.key, x.value as v1 FROM SRC x where x.key = 20 CLUSTER BY key
@@ -352,11 +352,11 @@
PREHOOK: query: SELECT x.key, x.value as v1 FROM SRC x where x.key = 20 CLUSTER BY key
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/327824513/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-54_276_5516587198210822350/10000
POSTHOOK: query: SELECT x.key, x.value as v1 FROM SRC x where x.key = 20 CLUSTER BY key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/327824513/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-54_276_5516587198210822350/10000
20 val_20
PREHOOK: query: EXPLAIN
SELECT x.key, x.value as v1 FROM SRC x where x.key = 20 CLUSTER BY x.key
@@ -424,11 +424,11 @@
PREHOOK: query: SELECT x.key, x.value as v1 FROM SRC x where x.key = 20 CLUSTER BY x.key
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/714056259/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-59_582_505878830166802854/10000
POSTHOOK: query: SELECT x.key, x.value as v1 FROM SRC x where x.key = 20 CLUSTER BY x.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/714056259/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-17-59_582_505878830166802854/10000
20 val_20
PREHOOK: query: EXPLAIN
SELECT x.key, x.value as v1 FROM SRC x where x.key = 20 CLUSTER BY v1
@@ -496,11 +496,11 @@
PREHOOK: query: SELECT x.key, x.value as v1 FROM SRC x where x.key = 20 CLUSTER BY v1
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/407934629/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-04_613_2939682790554351791/10000
POSTHOOK: query: SELECT x.key, x.value as v1 FROM SRC x where x.key = 20 CLUSTER BY v1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/407934629/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-04_613_2939682790554351791/10000
20 val_20
PREHOOK: query: EXPLAIN
SELECT y.* from (SELECT x.* FROM SRC x CLUSTER BY x.key) y where y.key = 20
@@ -571,11 +571,11 @@
PREHOOK: query: SELECT y.* from (SELECT x.* FROM SRC x CLUSTER BY x.key) y where y.key = 20
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/1402413625/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-10_209_6387108226324326472/10000
POSTHOOK: query: SELECT y.* from (SELECT x.* FROM SRC x CLUSTER BY x.key) y where y.key = 20
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/1402413625/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-10_209_6387108226324326472/10000
20 val_20
PREHOOK: query: EXPLAIN
SELECT x.key, x.value as v1, y.key FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY v1
@@ -595,21 +595,6 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- y
- TableScan
- alias: y
- Reduce Output Operator
- key expressions:
- expr: key
- type: string
- sort order: +
- Map-reduce partition columns:
- expr: key
- type: string
- tag: 1
- value expressions:
- expr: key
- type: string
x
TableScan
alias: x
@@ -631,6 +616,21 @@
type: string
expr: value
type: string
+ y
+ TableScan
+ alias: y
+ Reduce Output Operator
+ key expressions:
+ expr: key
+ type: string
+ sort order: +
+ Map-reduce partition columns:
+ expr: key
+ type: string
+ tag: 1
+ value expressions:
+ expr: key
+ type: string
Reduce Operator Tree:
Join Operator
condition map:
@@ -663,7 +663,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/1916601962/10002
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-16_334_2245182201779291081/10002
Reduce Output Operator
key expressions:
expr: _col1
@@ -697,11 +697,11 @@
PREHOOK: query: SELECT x.key, x.value as v1, y.key FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY v1
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/700452873/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-16_594_5072146069194835371/10000
POSTHOOK: query: SELECT x.key, x.value as v1, y.key FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY v1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/700452873/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-16_594_5072146069194835371/10000
20 val_20 20
PREHOOK: query: EXPLAIN
SELECT x.key, x.value as v1, y.* FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY v1
@@ -721,23 +721,6 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- y
- TableScan
- alias: y
- Reduce Output Operator
- key expressions:
- expr: key
- type: string
- sort order: +
- Map-reduce partition columns:
- expr: key
- type: string
- tag: 1
- value expressions:
- expr: key
- type: string
- expr: value
- type: string
x
TableScan
alias: x
@@ -759,6 +742,23 @@
type: string
expr: value
type: string
+ y
+ TableScan
+ alias: y
+ Reduce Output Operator
+ key expressions:
+ expr: key
+ type: string
+ sort order: +
+ Map-reduce partition columns:
+ expr: key
+ type: string
+ tag: 1
+ value expressions:
+ expr: key
+ type: string
+ expr: value
+ type: string
Reduce Operator Tree:
Join Operator
condition map:
@@ -793,7 +793,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/2139418033/10002
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-27_948_1334626995431478594/10002
Reduce Output Operator
key expressions:
expr: _col1
@@ -829,11 +829,11 @@
PREHOOK: query: SELECT x.key, x.value as v1, y.* FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY v1
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/1827492210/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-28_074_5978892602968822143/10000
POSTHOOK: query: SELECT x.key, x.value as v1, y.* FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY v1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/1827492210/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-28_074_5978892602968822143/10000
20 val_20 20 val_20
PREHOOK: query: EXPLAIN
SELECT x.key, x.value as v1, y.* FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY x.key
@@ -853,23 +853,6 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- y
- TableScan
- alias: y
- Reduce Output Operator
- key expressions:
- expr: key
- type: string
- sort order: +
- Map-reduce partition columns:
- expr: key
- type: string
- tag: 1
- value expressions:
- expr: key
- type: string
- expr: value
- type: string
x
TableScan
alias: x
@@ -891,6 +874,23 @@
type: string
expr: value
type: string
+ y
+ TableScan
+ alias: y
+ Reduce Output Operator
+ key expressions:
+ expr: key
+ type: string
+ sort order: +
+ Map-reduce partition columns:
+ expr: key
+ type: string
+ tag: 1
+ value expressions:
+ expr: key
+ type: string
+ expr: value
+ type: string
Reduce Operator Tree:
Join Operator
condition map:
@@ -925,7 +925,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/1673626041/10002
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-38_372_3122728094689915036/10002
Reduce Output Operator
key expressions:
expr: _col0
@@ -961,11 +961,11 @@
PREHOOK: query: SELECT x.key, x.value as v1, y.* FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY x.key
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/1237573825/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-38_466_4207748081161799368/10000
POSTHOOK: query: SELECT x.key, x.value as v1, y.* FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY x.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/1237573825/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-38_466_4207748081161799368/10000
20 val_20 20 val_20
PREHOOK: query: EXPLAIN
SELECT x.key, x.value as v1, y.key as yk FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY key
@@ -985,21 +985,6 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- y
- TableScan
- alias: y
- Reduce Output Operator
- key expressions:
- expr: key
- type: string
- sort order: +
- Map-reduce partition columns:
- expr: key
- type: string
- tag: 1
- value expressions:
- expr: key
- type: string
x
TableScan
alias: x
@@ -1021,6 +1006,21 @@
type: string
expr: value
type: string
+ y
+ TableScan
+ alias: y
+ Reduce Output Operator
+ key expressions:
+ expr: key
+ type: string
+ sort order: +
+ Map-reduce partition columns:
+ expr: key
+ type: string
+ tag: 1
+ value expressions:
+ expr: key
+ type: string
Reduce Operator Tree:
Join Operator
condition map:
@@ -1053,7 +1053,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/1226988138/10002
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-47_950_4989736117672479003/10002
Reduce Output Operator
key expressions:
expr: _col0
@@ -1087,11 +1087,11 @@
PREHOOK: query: SELECT x.key, x.value as v1, y.key as yk FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY key
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/302591922/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-48_073_5743203719731796760/10000
POSTHOOK: query: SELECT x.key, x.value as v1, y.key as yk FROM SRC x JOIN SRC y ON (x.key = y.key) where x.key = 20 CLUSTER BY key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/302591922/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-48_073_5743203719731796760/10000
20 val_20 20
PREHOOK: query: EXPLAIN
SELECT unioninput.*
@@ -1225,7 +1225,7 @@
CLUSTER BY unioninput.key
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/205488295/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-57_574_7730562904986793585/10000
POSTHOOK: query: SELECT unioninput.*
FROM (
FROM src select src.key, src.value WHERE src.key < 100
@@ -1235,7 +1235,7 @@
CLUSTER BY unioninput.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/205488295/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-18-57_574_7730562904986793585/10000
0 val_0
0 val_0
0 val_0
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/ctas.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/ctas.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/ctas.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/ctas.q.out Sun Feb 14 02:31:40 2010
@@ -26,11 +26,11 @@
PREHOOK: query: select * from nzhang_Tmp
PREHOOK: type: QUERY
PREHOOK: Input: default@nzhang_tmp
-PREHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/817024620/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-17-20_613_7362872865462991076/10000
POSTHOOK: query: select * from nzhang_Tmp
POSTHOOK: type: QUERY
POSTHOOK: Input: default@nzhang_tmp
-POSTHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/817024620/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-17-20_613_7362872865462991076/10000
PREHOOK: query: explain create table nzhang_CTAS1 as select key k, value from src sort by k, value limit 10
PREHOOK: type: CREATETABLE
POSTHOOK: query: explain create table nzhang_CTAS1 as select key k, value from src sort by k, value limit 10
@@ -84,7 +84,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/2138410252/10002
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-17-20_644_5490073691155494254/10002
Reduce Output Operator
key expressions:
expr: _col0
@@ -112,7 +112,7 @@
Move Operator
files:
hdfs directory: true
- destination: file:///data/users/jsichi/open/hive-trunk/ql/../build/ql/test/data/warehouse/nzhang_ctas1
+ destination: file:///data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/nzhang_ctas1
Stage: Stage-3
Create Table Operator:
@@ -136,11 +136,11 @@
PREHOOK: query: select * from nzhang_CTAS1
PREHOOK: type: QUERY
PREHOOK: Input: default@nzhang_ctas1
-PREHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/710958471/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-17-30_158_5988163105550483771/10000
POSTHOOK: query: select * from nzhang_CTAS1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@nzhang_ctas1
-POSTHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/710958471/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-17-30_158_5988163105550483771/10000
0 val_0
0 val_0
0 val_0
@@ -204,7 +204,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/305880921/10002
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-17-30_194_8489626264254862078/10002
Reduce Output Operator
key expressions:
expr: _col0
@@ -232,7 +232,7 @@
Move Operator
files:
hdfs directory: true
- destination: file:///data/users/jsichi/open/hive-trunk/ql/../build/ql/test/data/warehouse/nzhang_ctas2
+ destination: file:///data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/nzhang_ctas2
Stage: Stage-3
Create Table Operator:
@@ -256,11 +256,11 @@
PREHOOK: query: select * from nzhang_ctas2
PREHOOK: type: QUERY
PREHOOK: Input: default@nzhang_ctas2
-PREHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1731188757/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-17-39_797_5455952305307754482/10000
POSTHOOK: query: select * from nzhang_ctas2
POSTHOOK: type: QUERY
POSTHOOK: Input: default@nzhang_ctas2
-POSTHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1731188757/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-17-39_797_5455952305307754482/10000
0 val_0
0 val_0
0 val_0
@@ -324,7 +324,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/564962024/10002
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-17-39_831_4392155595765824679/10002
Reduce Output Operator
key expressions:
expr: _col0
@@ -352,7 +352,7 @@
Move Operator
files:
hdfs directory: true
- destination: file:///data/users/jsichi/open/hive-trunk/ql/../build/ql/test/data/warehouse/nzhang_ctas3
+ destination: file:///data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/nzhang_ctas3
Stage: Stage-3
Create Table Operator:
@@ -377,11 +377,11 @@
PREHOOK: query: select * from nzhang_ctas3
PREHOOK: type: QUERY
PREHOOK: Input: default@nzhang_ctas3
-PREHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/535377929/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-17-51_297_1404107458733070450/10000
POSTHOOK: query: select * from nzhang_ctas3
POSTHOOK: type: QUERY
POSTHOOK: Input: default@nzhang_ctas3
-POSTHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/535377929/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-17-51_297_1404107458733070450/10000
0.0 val_0_con
0.0 val_0_con
0.0 val_0_con
@@ -410,11 +410,11 @@
PREHOOK: query: select * from nzhang_ctas3
PREHOOK: type: QUERY
PREHOOK: Input: default@nzhang_ctas3
-PREHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1725800016/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-17-51_421_411454597887039980/10000
POSTHOOK: query: select * from nzhang_ctas3
POSTHOOK: type: QUERY
POSTHOOK: Input: default@nzhang_ctas3
-POSTHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1725800016/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-17-51_421_411454597887039980/10000
0.0 val_0_con
0.0 val_0_con
0.0 val_0_con
@@ -478,7 +478,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/633210786/10002
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-17-51_456_7736262908389950265/10002
Reduce Output Operator
key expressions:
expr: _col0
@@ -506,7 +506,7 @@
Move Operator
files:
hdfs directory: true
- destination: file:///data/users/jsichi/open/hive-trunk/ql/../build/ql/test/data/warehouse/nzhang_ctas4
+ destination: file:///data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/nzhang_ctas4
Stage: Stage-3
Create Table Operator:
@@ -531,11 +531,11 @@
PREHOOK: query: select * from nzhang_ctas4
PREHOOK: type: QUERY
PREHOOK: Input: default@nzhang_ctas4
-PREHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1216389488/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-18-04_556_491115919333585412/10000
POSTHOOK: query: select * from nzhang_ctas4
POSTHOOK: type: QUERY
POSTHOOK: Input: default@nzhang_ctas4
-POSTHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1216389488/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-18-04_556_491115919333585412/10000
0 val_0
0 val_0
0 val_0
@@ -588,41 +588,41 @@
type: string
Needs Tagging: false
Path -> Alias:
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src [src]
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/src [src]
Path -> Partition:
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/src
Partition
base file name: src
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name src
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/src
+ name src
serialization.ddl struct src { string key, string value}
serialization.format 1
- columns key,value
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
- transient_lastDdlTime 1262135512
+ transient_lastDdlTime 1266041840
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name src
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/src
+ name src
serialization.ddl struct src { string key, string value}
serialization.format 1
- columns key,value
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
- transient_lastDdlTime 1262135512
+ transient_lastDdlTime 1266041840
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: src
name: src
@@ -632,7 +632,7 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1017044863/10002
+ directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-18-04_591_5195160466418251956/10002
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -644,7 +644,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1017044863/10002
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-18-04_591_5195160466418251956/10002
Reduce Output Operator
key expressions:
expr: _col0
@@ -660,9 +660,9 @@
type: string
Needs Tagging: false
Path -> Alias:
- file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1017044863/10002 [file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1017044863/10002]
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-18-04_591_5195160466418251956/10002 [file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-18-04_591_5195160466418251956/10002]
Path -> Partition:
- file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1017044863/10002
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-18-04_591_5195160466418251956/10002
Partition
base file name: 10002
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -684,24 +684,24 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1017044863/10001
+ directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-18-04_591_5195160466418251956/10001
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ columns _col0,_col1
+ columns.types string:string
field.delim ,
line.delim
- columns _col0,_col1
serialization.format ,
- columns.types string:string
Stage: Stage-0
Move Operator
files:
hdfs directory: true
- source: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1017044863/10001
- destination: file:///data/users/jsichi/open/hive-trunk/ql/../build/ql/test/data/warehouse/nzhang_ctas5
+ source: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-18-04_591_5195160466418251956/10001
+ destination: file:///data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/nzhang_ctas5
Stage: Stage-3
Create Table Operator:
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby_map_ppr.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby_map_ppr.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby_map_ppr.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby_map_ppr.q.out Sun Feb 14 02:31:40 2010
@@ -78,10 +78,10 @@
type: double
Needs Tagging: false
Path -> Alias:
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [src]
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [src]
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [src]
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [src]
Path -> Partition:
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
Partition
base file name: hr=11
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -90,39 +90,39 @@
ds 2008-04-08
hr 11
properties:
- name srcpart
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/srcpart
+ name srcpart
+ partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
- columns key,value
- partition_columns ds/hr
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart
- transient_lastDdlTime 1262135893
+ transient_lastDdlTime 1266042054
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name srcpart
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/srcpart
+ name srcpart
+ partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
- columns key,value
- partition_columns ds/hr
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart
- transient_lastDdlTime 1262135893
+ transient_lastDdlTime 1266042054
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
name: srcpart
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
Partition
base file name: hr=12
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -131,35 +131,35 @@
ds 2008-04-08
hr 12
properties:
- name srcpart
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/srcpart
+ name srcpart
+ partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
- columns key,value
- partition_columns ds/hr
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart
- transient_lastDdlTime 1262135893
+ transient_lastDdlTime 1266042054
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name srcpart
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/srcpart
+ name srcpart
+ partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
- columns key,value
- partition_columns ds/hr
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart
- transient_lastDdlTime 1262135893
+ transient_lastDdlTime 1266042054
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
name: srcpart
@@ -195,22 +195,22 @@
File Output Operator
compressed: false
GlobalTableId: 1
- directory: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1077000950/10000
+ directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-20-56_866_3346083160698669953/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name dest1
+ bucket_count -1
+ columns key,c1,c2
columns.types string:int:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/dest1
+ name dest1
serialization.ddl struct dest1 { string key, i32 c1, string c2}
serialization.format 1
- columns key,c1,c2
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/dest1
- transient_lastDdlTime 1262135895
+ transient_lastDdlTime 1266042056
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest1
@@ -218,25 +218,25 @@
Move Operator
tables:
replace: true
- source: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1077000950/10000
+ source: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-20-56_866_3346083160698669953/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name dest1
+ bucket_count -1
+ columns key,c1,c2
columns.types string:int:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/dest1
+ name dest1
serialization.ddl struct dest1 { string key, i32 c1, string c2}
serialization.format 1
- columns key,c1,c2
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/dest1
- transient_lastDdlTime 1262135895
+ transient_lastDdlTime 1266042056
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest1
- tmp directory: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1077000950/10001
+ tmp directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-20-56_866_3346083160698669953/10001
PREHOOK: query: FROM srcpart src
@@ -260,11 +260,11 @@
PREHOOK: query: SELECT dest1.* FROM dest1
PREHOOK: type: QUERY
PREHOOK: Input: default@dest1
-PREHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1579733451/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-21-02_407_3581752610986474753/10000
POSTHOOK: query: SELECT dest1.* FROM dest1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest1
-POSTHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1579733451/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-21-02_407_3581752610986474753/10000
0 1 00.0
1 71 132828.0
2 69 251142.0
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby_ppr.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby_ppr.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby_ppr.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby_ppr.q.out Sun Feb 14 02:31:40 2010
@@ -61,10 +61,10 @@
tag: -1
Needs Tagging: false
Path -> Alias:
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [src]
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [src]
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [src]
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [src]
Path -> Partition:
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
Partition
base file name: hr=11
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -73,39 +73,39 @@
ds 2008-04-08
hr 11
properties:
- name srcpart
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart
+ name srcpart
+ partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
- columns key,value
- partition_columns ds/hr
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart
- transient_lastDdlTime 1262135908
+ transient_lastDdlTime 1266042341
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name srcpart
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart
+ name srcpart
+ partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
- columns key,value
- partition_columns ds/hr
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart
- transient_lastDdlTime 1262135908
+ transient_lastDdlTime 1266042341
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
name: srcpart
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
Partition
base file name: hr=12
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -114,35 +114,35 @@
ds 2008-04-08
hr 12
properties:
- name srcpart
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart
+ name srcpart
+ partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
- columns key,value
- partition_columns ds/hr
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart
- transient_lastDdlTime 1262135908
+ transient_lastDdlTime 1266042341
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name srcpart
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart
+ name srcpart
+ partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
- columns key,value
- partition_columns ds/hr
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart
- transient_lastDdlTime 1262135908
+ transient_lastDdlTime 1266042341
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
name: srcpart
@@ -178,22 +178,22 @@
File Output Operator
compressed: false
GlobalTableId: 1
- directory: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1636998037/10000
+ directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-25-43_755_1556731248680627580/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name dest1
+ bucket_count -1
+ columns key,c1,c2
columns.types string:int:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest1
+ name dest1
serialization.ddl struct dest1 { string key, i32 c1, string c2}
serialization.format 1
- columns key,c1,c2
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/dest1
- transient_lastDdlTime 1262135909
+ transient_lastDdlTime 1266042343
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest1
@@ -201,25 +201,25 @@
Move Operator
tables:
replace: true
- source: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1636998037/10000
+ source: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-25-43_755_1556731248680627580/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name dest1
+ bucket_count -1
+ columns key,c1,c2
columns.types string:int:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest1
+ name dest1
serialization.ddl struct dest1 { string key, i32 c1, string c2}
serialization.format 1
- columns key,c1,c2
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/dest1
- transient_lastDdlTime 1262135909
+ transient_lastDdlTime 1266042343
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest1
- tmp directory: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1636998037/10001
+ tmp directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-25-43_755_1556731248680627580/10001
PREHOOK: query: FROM srcpart src
@@ -243,11 +243,11 @@
PREHOOK: query: SELECT dest1.* FROM dest1
PREHOOK: type: QUERY
PREHOOK: Input: default@dest1
-PREHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1831892895/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-25-50_172_6197779811255758201/10000
POSTHOOK: query: SELECT dest1.* FROM dest1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest1
-POSTHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1831892895/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-25-50_172_6197779811255758201/10000
0 1 00.0
1 71 132828.0
2 69 251142.0