You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by nz...@apache.org on 2010/02/14 03:31:44 UTC
svn commit: r909965 [9/15] - in /hadoop/hive/trunk: ./
contrib/src/test/results/clientnegative/
contrib/src/test/results/clientpositive/
ql/src/java/org/apache/hadoop/hive/ql/exec/
ql/src/test/results/clientpositive/
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_join3.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_join3.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_join3.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_join3.q.out Sun Feb 14 02:31:40 2010
@@ -33,38 +33,6 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- src2:src
- TableScan
- alias: src
- Filter Operator
- predicate:
- expr: ((key <> '12') and (key <> '4'))
- type: boolean
- Filter Operator
- predicate:
- expr: (key <> '12')
- type: boolean
- Select Operator
- expressions:
- expr: key
- type: string
- expr: value
- type: string
- outputColumnNames: _col0, _col1
- Reduce Output Operator
- key expressions:
- expr: _col0
- type: string
- sort order: +
- Map-reduce partition columns:
- expr: _col0
- type: string
- tag: 1
- value expressions:
- expr: _col0
- type: string
- expr: _col1
- type: string
src1:src
TableScan
alias: src
@@ -101,6 +69,38 @@
type: string
expr: _col1
type: string
+ src2:src
+ TableScan
+ alias: src
+ Filter Operator
+ predicate:
+ expr: ((key <> '12') and (key <> '4'))
+ type: boolean
+ Filter Operator
+ predicate:
+ expr: (key <> '12')
+ type: boolean
+ Select Operator
+ expressions:
+ expr: key
+ type: string
+ expr: value
+ type: string
+ outputColumnNames: _col0, _col1
+ Reduce Output Operator
+ key expressions:
+ expr: _col0
+ type: string
+ sort order: +
+ Map-reduce partition columns:
+ expr: _col0
+ type: string
+ tag: 1
+ value expressions:
+ expr: _col0
+ type: string
+ expr: _col1
+ type: string
src3:src
TableScan
alias: src
@@ -175,7 +175,7 @@
WHERE src1.c1 > '0' and (src1.c2 <> 'val_500' or src1.c1 > '1') and (src2.c3 > '10' or src1.c1 <> '10') and (src2.c3 <> '4') and (src3.c5 <> '1')
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/543319193/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_3/build/ql/scratchdir/hive_2010-02-12_22-18-31_804_5890089031375913660/10000
POSTHOOK: query: SELECT src1.c1, src2.c4
FROM
(SELECT src.key as c1, src.value as c2 from src where src.key <> '11' ) src1
@@ -188,7 +188,7 @@
WHERE src1.c1 > '0' and (src1.c2 <> 'val_500' or src1.c1 > '1') and (src2.c3 > '10' or src1.c1 <> '10') and (src2.c3 <> '4') and (src3.c5 <> '1')
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/543319193/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_3/build/ql/scratchdir/hive_2010-02-12_22-18-31_804_5890089031375913660/10000
100 val_100
100 val_100
100 val_100
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_multi_insert.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_multi_insert.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_multi_insert.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_multi_insert.q.out Sun Feb 14 02:31:40 2010
@@ -53,9 +53,9 @@
Stage: Stage-4
Map Reduce
Alias -> Map Operator Tree:
- b
+ a
TableScan
- alias: b
+ alias: a
Reduce Output Operator
key expressions:
expr: key
@@ -64,10 +64,15 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 1
- a
+ tag: 0
+ value expressions:
+ expr: key
+ type: string
+ expr: value
+ type: string
+ b
TableScan
- alias: a
+ alias: b
Reduce Output Operator
key expressions:
expr: key
@@ -76,12 +81,7 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 0
- value expressions:
- expr: key
- type: string
- expr: value
- type: string
+ tag: 1
Reduce Operator Tree:
Join Operator
condition map:
@@ -246,11 +246,11 @@
PREHOOK: query: SELECT mi1.* FROM mi1
PREHOOK: type: QUERY
PREHOOK: Input: default@mi1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/775375392/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-34-41_902_6905591894666460215/10000
POSTHOOK: query: SELECT mi1.* FROM mi1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@mi1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/775375392/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-34-41_902_6905591894666460215/10000
0 val_0
0 val_0
0 val_0
@@ -402,11 +402,11 @@
PREHOOK: query: SELECT mi2.* FROM mi2
PREHOOK: type: QUERY
PREHOOK: Input: default@mi2
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/106849464/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-34-41_936_9202557721170644347/10000
POSTHOOK: query: SELECT mi2.* FROM mi2
POSTHOOK: type: QUERY
POSTHOOK: Input: default@mi2
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/106849464/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-34-41_936_9202557721170644347/10000
100 val_100
100 val_100
100 val_100
@@ -621,11 +621,11 @@
PREHOOK: query: SELECT mi3.* FROM mi3
PREHOOK: type: QUERY
PREHOOK: Input: default@mi3@ds=2008-04-08/hr=12
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1706731089/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-34-41_968_153693361750093127/10000
POSTHOOK: query: SELECT mi3.* FROM mi3
POSTHOOK: type: QUERY
POSTHOOK: Input: default@mi3@ds=2008-04-08/hr=12
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1706731089/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-34-41_968_153693361750093127/10000
200 2008-04-08 12
200 2008-04-08 12
200 2008-04-08 12
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join1.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join1.q.out Sun Feb 14 02:31:40 2010
@@ -27,23 +27,6 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- b
- TableScan
- alias: b
- Reduce Output Operator
- key expressions:
- expr: key
- type: string
- sort order: +
- Map-reduce partition columns:
- expr: key
- type: string
- tag: 1
- value expressions:
- expr: key
- type: string
- expr: value
- type: string
a
TableScan
alias: a
@@ -65,6 +48,23 @@
type: string
expr: value
type: string
+ b
+ TableScan
+ alias: b
+ Reduce Output Operator
+ key expressions:
+ expr: key
+ type: string
+ sort order: +
+ Map-reduce partition columns:
+ expr: key
+ type: string
+ tag: 1
+ value expressions:
+ expr: key
+ type: string
+ expr: value
+ type: string
Reduce Operator Tree:
Join Operator
condition map:
@@ -110,7 +110,7 @@
WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/244410137/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_3/build/ql/scratchdir/hive_2010-02-12_22-18-41_861_2988494112154470083/10000
POSTHOOK: query: FROM
src a
LEFT OUTER JOIN
@@ -120,7 +120,7 @@
WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/244410137/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_3/build/ql/scratchdir/hive_2010-02-12_22-18-41_861_2988494112154470083/10000
17 val_17 17 val_17
18 val_18 18 val_18
18 val_18 18 val_18
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join2.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join2.q.out Sun Feb 14 02:31:40 2010
@@ -27,6 +27,23 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
+ a
+ TableScan
+ alias: a
+ Reduce Output Operator
+ key expressions:
+ expr: key
+ type: string
+ sort order: +
+ Map-reduce partition columns:
+ expr: key
+ type: string
+ tag: 0
+ value expressions:
+ expr: key
+ type: string
+ expr: value
+ type: string
b
TableScan
alias: b
@@ -48,23 +65,6 @@
type: string
expr: value
type: string
- a
- TableScan
- alias: a
- Reduce Output Operator
- key expressions:
- expr: key
- type: string
- sort order: +
- Map-reduce partition columns:
- expr: key
- type: string
- tag: 0
- value expressions:
- expr: key
- type: string
- expr: value
- type: string
Reduce Operator Tree:
Join Operator
condition map:
@@ -110,7 +110,7 @@
WHERE a.key > '10' AND a.key < '20' AND b.key > '15' AND b.key < '25'
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/1607355712/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_3/build/ql/scratchdir/hive_2010-02-12_22-18-49_166_7655816003584089890/10000
POSTHOOK: query: FROM
src a
RIGHT OUTER JOIN
@@ -120,7 +120,7 @@
WHERE a.key > '10' AND a.key < '20' AND b.key > '15' AND b.key < '25'
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/1607355712/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_3/build/ql/scratchdir/hive_2010-02-12_22-18-49_166_7655816003584089890/10000
150 val_150 150 val_150
152 val_152 152 val_152
152 val_152 152 val_152
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join3.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join3.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join3.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join3.q.out Sun Feb 14 02:31:40 2010
@@ -27,9 +27,9 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- b
+ a
TableScan
- alias: b
+ alias: a
Reduce Output Operator
key expressions:
expr: key
@@ -38,15 +38,15 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 1
+ tag: 0
value expressions:
expr: key
type: string
expr: value
type: string
- a
+ b
TableScan
- alias: a
+ alias: b
Reduce Output Operator
key expressions:
expr: key
@@ -55,7 +55,7 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 0
+ tag: 1
value expressions:
expr: key
type: string
@@ -106,7 +106,7 @@
WHERE a.key > '10' AND a.key < '20' AND b.key > '15' AND b.key < '25'
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/794158250/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_3/build/ql/scratchdir/hive_2010-02-12_22-18-58_300_1623871517298833290/10000
POSTHOOK: query: FROM
src a
FULL OUTER JOIN
@@ -116,7 +116,7 @@
WHERE a.key > '10' AND a.key < '20' AND b.key > '15' AND b.key < '25'
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/794158250/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_3/build/ql/scratchdir/hive_2010-02-12_22-18-58_300_1623871517298833290/10000
150 val_150 150 val_150
152 val_152 152 val_152
152 val_152 152 val_152
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join4.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join4.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join4.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join4.q.out Sun Feb 14 02:31:40 2010
@@ -33,9 +33,9 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- b
+ a
TableScan
- alias: b
+ alias: a
Reduce Output Operator
key expressions:
expr: key
@@ -44,15 +44,15 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 1
+ tag: 0
value expressions:
expr: key
type: string
expr: value
type: string
- c
+ b
TableScan
- alias: c
+ alias: b
Reduce Output Operator
key expressions:
expr: key
@@ -61,13 +61,15 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 2
+ tag: 1
value expressions:
expr: key
type: string
- a
+ expr: value
+ type: string
+ c
TableScan
- alias: a
+ alias: c
Reduce Output Operator
key expressions:
expr: key
@@ -76,12 +78,10 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 0
+ tag: 2
value expressions:
expr: key
type: string
- expr: value
- type: string
Reduce Operator Tree:
Join Operator
condition map:
@@ -134,7 +134,7 @@
WHERE a.key > '10' AND a.key < '20' AND b.key > '15' AND b.key < '25' AND sqrt(c.key) <> 13
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/1320647695/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-31-48_596_2953182954930521327/10000
POSTHOOK: query: FROM
src a
LEFT OUTER JOIN
@@ -147,7 +147,7 @@
WHERE a.key > '10' AND a.key < '20' AND b.key > '15' AND b.key < '25' AND sqrt(c.key) <> 13
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/1320647695/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-31-48_596_2953182954930521327/10000
150 val_150 150 val_150 150
152 val_152 152 val_152 152
152 val_152 152 val_152 152
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_random.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_random.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_random.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_random.q.out Sun Feb 14 02:31:40 2010
@@ -27,6 +27,26 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
+ src1:src
+ TableScan
+ alias: src
+ Select Operator
+ expressions:
+ expr: key
+ type: string
+ outputColumnNames: _col0
+ Reduce Output Operator
+ key expressions:
+ expr: _col0
+ type: string
+ sort order: +
+ Map-reduce partition columns:
+ expr: _col0
+ type: string
+ tag: 0
+ value expressions:
+ expr: _col0
+ type: string
src2:src
TableScan
alias: src
@@ -57,26 +77,6 @@
value expressions:
expr: _col1
type: string
- src1:src
- TableScan
- alias: src
- Select Operator
- expressions:
- expr: key
- type: string
- outputColumnNames: _col0
- Reduce Output Operator
- key expressions:
- expr: _col0
- type: string
- sort order: +
- Map-reduce partition columns:
- expr: _col0
- type: string
- tag: 0
- value expressions:
- expr: _col0
- type: string
Reduce Operator Tree:
Join Operator
condition map:
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_udf_case.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_udf_case.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_udf_case.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_udf_case.q.out Sun Feb 14 02:31:40 2010
@@ -36,12 +36,12 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- b
+ a
TableScan
- alias: b
+ alias: a
Filter Operator
predicate:
- expr: (ds = '2008-04-08')
+ expr: ((ds = '2008-04-08') and CASE (key) WHEN ('27') THEN (true) WHEN ('38') THEN (false) ELSE (null) END)
type: boolean
Reduce Output Operator
key expressions:
@@ -51,7 +51,7 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 1
+ tag: 0
value expressions:
expr: key
type: string
@@ -61,12 +61,12 @@
type: string
expr: hr
type: string
- a
+ b
TableScan
- alias: a
+ alias: b
Filter Operator
predicate:
- expr: ((ds = '2008-04-08') and CASE (key) WHEN ('27') THEN (true) WHEN ('38') THEN (false) ELSE (null) END)
+ expr: (ds = '2008-04-08')
type: boolean
Reduce Output Operator
key expressions:
@@ -76,7 +76,7 @@
Map-reduce partition columns:
expr: key
type: string
- tag: 0
+ tag: 1
value expressions:
expr: key
type: string
@@ -128,7 +128,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/438623182/10002
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-31-56_689_3254720341447490427/10002
Reduce Output Operator
key expressions:
expr: _col0
@@ -194,7 +194,7 @@
PREHOOK: type: QUERY
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/2137110555/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-31-56_869_4629499201500113908/10000
POSTHOOK: query: SELECT *
FROM srcpart a JOIN srcpart b
ON a.key = b.key
@@ -209,7 +209,7 @@
POSTHOOK: type: QUERY
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/2137110555/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-31-56_869_4629499201500113908/10000
27 val_27 2008-04-08 11 27 val_27 2008-04-08 11
27 val_27 2008-04-08 11 27 val_27 2008-04-08 12
27 val_27 2008-04-08 12 27 val_27 2008-04-08 11
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner1.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner1.q.out Sun Feb 14 02:31:40 2010
@@ -33,51 +33,51 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1281920331/10001
+ directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-35-08_059_3516688999660848477/10001
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
columns _col0,_col1
- serialization.format 1
columns.types string:string
+ serialization.format 1
Needs Tagging: false
Path -> Alias:
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src [src]
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/src [src]
Path -> Partition:
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/src
Partition
base file name: src
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name src
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/src
+ name src
serialization.ddl struct src { string key, string value}
serialization.format 1
- columns key,value
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
- transient_lastDdlTime 1262137321
+ transient_lastDdlTime 1266042907
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name src
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/src
+ name src
serialization.ddl struct src { string key, string value}
serialization.format 1
- columns key,value
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
- transient_lastDdlTime 1262137321
+ transient_lastDdlTime 1266042907
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: src
name: src
@@ -90,11 +90,11 @@
PREHOOK: query: select * from src where rand(1) < 0.1
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1343937738/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-35-08_099_7501446808782656360/10000
POSTHOOK: query: select * from src where rand(1) < 0.1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1343937738/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-35-08_099_7501446808782656360/10000
409 val_409
429 val_429
209 val_209
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner2.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner2.q.out Sun Feb 14 02:31:40 2010
@@ -53,30 +53,30 @@
File Output Operator
compressed: false
GlobalTableId: 1
- directory: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/297465583/10002
+ directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-21_884_2937772000010325968/10002
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name tmptable
+ bucket_count -1
+ columns key,value,hr,ds
columns.types string:string:string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/tmptable
+ name tmptable
serialization.ddl struct tmptable { string key, string value, string hr, string ds}
serialization.format 1
- columns key,value,hr,ds
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/tmptable
- transient_lastDdlTime 1263595477
+ transient_lastDdlTime 1266042741
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: tmptable
Needs Tagging: false
Path -> Alias:
- file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [a]
- file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [a]
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [a]
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [a]
Path -> Partition:
- file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
Partition
base file name: hr=11
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -85,39 +85,39 @@
ds 2008-04-08
hr 11
properties:
- name srcpart
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/srcpart
+ name srcpart
+ partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
- columns key,value
- partition_columns ds/hr
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/srcpart
- transient_lastDdlTime 1263595476
+ transient_lastDdlTime 1266042740
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name srcpart
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/srcpart
+ name srcpart
+ partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
- columns key,value
- partition_columns ds/hr
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/srcpart
- transient_lastDdlTime 1263595476
+ transient_lastDdlTime 1266042740
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
name: srcpart
- file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
Partition
base file name: hr=12
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -126,35 +126,35 @@
ds 2008-04-08
hr 12
properties:
- name srcpart
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/srcpart
+ name srcpart
+ partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
- columns key,value
- partition_columns ds/hr
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/srcpart
- transient_lastDdlTime 1263595476
+ transient_lastDdlTime 1266042740
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name srcpart
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/srcpart
+ name srcpart
+ partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
- columns key,value
- partition_columns ds/hr
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/srcpart
- transient_lastDdlTime 1263595476
+ transient_lastDdlTime 1266042740
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
name: srcpart
@@ -166,37 +166,37 @@
Move Operator
files:
hdfs directory: true
- source: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/297465583/10002
- destination: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/567762797/10000
+ source: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-21_884_2937772000010325968/10002
+ destination: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-21_884_2937772000010325968/10000
Stage: Stage-0
Move Operator
tables:
replace: true
- source: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/567762797/10000
+ source: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-21_884_2937772000010325968/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name tmptable
+ bucket_count -1
+ columns key,value,hr,ds
columns.types string:string:string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/tmptable
+ name tmptable
serialization.ddl struct tmptable { string key, string value, string hr, string ds}
serialization.format 1
- columns key,value,hr,ds
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/tmptable
- transient_lastDdlTime 1263595477
+ transient_lastDdlTime 1266042741
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: tmptable
- tmp directory: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/567762797/10001
+ tmp directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-21_884_2937772000010325968/10001
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/297465583/10002
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-21_884_2937772000010325968/10002
Reduce Output Operator
sort order:
Map-reduce partition columns:
@@ -214,41 +214,41 @@
type: string
Needs Tagging: false
Path -> Alias:
- file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/297465583/10002 [file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/297465583/10002]
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-21_884_2937772000010325968/10002 [file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-21_884_2937772000010325968/10002]
Path -> Partition:
- file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/297465583/10002
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-21_884_2937772000010325968/10002
Partition
base file name: 10002
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name tmptable
+ bucket_count -1
+ columns key,value,hr,ds
columns.types string:string:string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/tmptable
+ name tmptable
serialization.ddl struct tmptable { string key, string value, string hr, string ds}
serialization.format 1
- columns key,value,hr,ds
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/tmptable
- transient_lastDdlTime 1263595477
+ transient_lastDdlTime 1266042741
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name tmptable
+ bucket_count -1
+ columns key,value,hr,ds
columns.types string:string:string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/tmptable
+ name tmptable
serialization.ddl struct tmptable { string key, string value, string hr, string ds}
serialization.format 1
- columns key,value,hr,ds
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/tmptable
- transient_lastDdlTime 1263595477
+ transient_lastDdlTime 1266042741
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: tmptable
name: tmptable
@@ -257,22 +257,22 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/567762797/10000
+ directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-21_884_2937772000010325968/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name tmptable
+ bucket_count -1
+ columns key,value,hr,ds
columns.types string:string:string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/tmptable
+ name tmptable
serialization.ddl struct tmptable { string key, string value, string hr, string ds}
serialization.format 1
- columns key,value,hr,ds
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- location file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/tmptable
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- transient_lastDdlTime 1263595477
+ transient_lastDdlTime 1266042741
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: tmptable
@@ -292,11 +292,11 @@
PREHOOK: query: select * from tmptable x sort by x.key,x.value,x.ds,x.hr
PREHOOK: type: QUERY
PREHOOK: Input: default@tmptable
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/543578106/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-32_891_7737984582274649304/10000
POSTHOOK: query: select * from tmptable x sort by x.key,x.value,x.ds,x.hr
POSTHOOK: type: QUERY
POSTHOOK: Input: default@tmptable
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/543578106/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-32_891_7737984582274649304/10000
103 val_103 2008-04-08 11
103 val_103 2008-04-08 12
133 val_133 2008-04-08 11
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner3.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner3.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner3.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner3.q.out Sun Feb 14 02:31:40 2010
@@ -39,19 +39,19 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/124088894/10001
+ directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-37-59_832_3788417974183008166/10001
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
columns _col0,_col1,_col2,_col3
- serialization.format 1
columns.types string:string:string:string
+ serialization.format 1
Needs Tagging: false
Path -> Alias:
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [a]
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [a]
Path -> Partition:
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
Partition
base file name: hr=12
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -60,35 +60,35 @@
ds 2008-04-08
hr 12
properties:
- name srcpart
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart
+ name srcpart
+ partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
- columns key,value
- partition_columns ds/hr
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart
- transient_lastDdlTime 1262137336
+ transient_lastDdlTime 1266043078
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name srcpart
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart
+ name srcpart
+ partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
- columns key,value
- partition_columns ds/hr
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart
- transient_lastDdlTime 1262137336
+ transient_lastDdlTime 1266043078
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
name: srcpart
@@ -101,11 +101,11 @@
PREHOOK: query: select a.* from srcpart a where rand(1) < 0.1 and a.ds = '2008-04-08' and not(key > 50 or key < 10) and a.hr like '%2'
PREHOOK: type: QUERY
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-PREHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1944369228/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-37-59_919_2841312693651221074/10000
POSTHOOK: query: select a.* from srcpart a where rand(1) < 0.1 and a.ds = '2008-04-08' and not(key > 50 or key < 10) and a.hr like '%2'
POSTHOOK: type: QUERY
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-POSTHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1944369228/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-37-59_919_2841312693651221074/10000
42 val_42 2008-04-08 12
44 val_44 2008-04-08 12
26 val_26 2008-04-08 12
@@ -155,19 +155,19 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1956461194/10001
+ directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-38-03_773_7561053502233744490/10001
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
columns _col0,_col1,_col2,_col3
- serialization.format 1
columns.types string:string:string:string
+ serialization.format 1
Needs Tagging: false
Path -> Alias:
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [a]
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [a]
Path -> Partition:
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
Partition
base file name: hr=12
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -176,35 +176,35 @@
ds 2008-04-08
hr 12
properties:
- name srcpart
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart
+ name srcpart
+ partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
- columns key,value
- partition_columns ds/hr
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart
- transient_lastDdlTime 1262137336
+ transient_lastDdlTime 1266043078
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name srcpart
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart
+ name srcpart
+ partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
- columns key,value
- partition_columns ds/hr
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart
- transient_lastDdlTime 1262137336
+ transient_lastDdlTime 1266043078
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
name: srcpart
@@ -217,11 +217,11 @@
PREHOOK: query: select a.* from srcpart a where a.ds = '2008-04-08' and not(key > 50 or key < 10) and a.hr like '%2'
PREHOOK: type: QUERY
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-PREHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1681163814/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-38-03_865_149659006171751209/10000
POSTHOOK: query: select a.* from srcpart a where a.ds = '2008-04-08' and not(key > 50 or key < 10) and a.hr like '%2'
POSTHOOK: type: QUERY
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-POSTHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1681163814/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-38-03_865_149659006171751209/10000
27 val_27 2008-04-08 12
37 val_37 2008-04-08 12
15 val_15 2008-04-08 12
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/regex_col.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/regex_col.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/regex_col.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/regex_col.q.out Sun Feb 14 02:31:40 2010
@@ -113,9 +113,9 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- b
+ a
TableScan
- alias: b
+ alias: a
Reduce Output Operator
key expressions:
expr: key
@@ -128,15 +128,15 @@
type: string
expr: value
type: string
- tag: 1
+ tag: 0
value expressions:
expr: ds
type: string
expr: hr
type: string
- a
+ b
TableScan
- alias: a
+ alias: b
Reduce Output Operator
key expressions:
expr: key
@@ -149,7 +149,7 @@
type: string
expr: value
type: string
- tag: 0
+ tag: 1
value expressions:
expr: ds
type: string
@@ -209,31 +209,6 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- b
- TableScan
- alias: b
- Reduce Output Operator
- key expressions:
- expr: key
- type: string
- expr: hr
- type: string
- expr: ds
- type: string
- sort order: +++
- Map-reduce partition columns:
- expr: key
- type: string
- expr: hr
- type: string
- expr: ds
- type: string
- tag: 1
- value expressions:
- expr: ds
- type: string
- expr: hr
- type: string
a
TableScan
alias: a
@@ -262,6 +237,31 @@
expr: ds
type: string
tag: 0
+ b
+ TableScan
+ alias: b
+ Reduce Output Operator
+ key expressions:
+ expr: key
+ type: string
+ expr: hr
+ type: string
+ expr: ds
+ type: string
+ sort order: +++
+ Map-reduce partition columns:
+ expr: key
+ type: string
+ expr: hr
+ type: string
+ expr: ds
+ type: string
+ tag: 1
+ value expressions:
+ expr: ds
+ type: string
+ expr: hr
+ type: string
Reduce Operator Tree:
Join Operator
condition map:
@@ -288,7 +288,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/832608785/10002
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-36-24_399_2804249481999873947/10002
Reduce Output Operator
key expressions:
expr: _col0
@@ -324,7 +324,7 @@
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1495752848/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-36-24_582_4644469832031285008/10000
POSTHOOK: query: SELECT b.`..` FROM srcpart a JOIN srcpart b
ON a.key = b.key AND a.hr = b.hr AND a.ds = b.ds AND a.key = 103
ORDER BY ds, hr
@@ -333,7 +333,7 @@
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1495752848/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-36-24_582_4644469832031285008/10000
2008-04-08 11
2008-04-08 11
2008-04-08 11
@@ -526,14 +526,14 @@
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/316243993/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-36-36_271_5109114395556786366/10000
POSTHOOK: query: SELECT `(ds|hr)?+.+` FROM srcpart ORDER BY key, value LIMIT 10
POSTHOOK: type: QUERY
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/316243993/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-36-36_271_5109114395556786366/10000
0 val_0
0 val_0
0 val_0
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/regexp_extract.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/regexp_extract.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/regexp_extract.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/regexp_extract.q.out Sun Feb 14 02:31:40 2010
@@ -47,10 +47,10 @@
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- field.delim 9
columns _col0,_col1
- serialization.format 9
columns.types string,string
+ field.delim 9
+ serialization.format 9
serialization.last.column.takes.rest true
Reduce Output Operator
key expressions:
@@ -68,41 +68,41 @@
type: string
Needs Tagging: false
Path -> Alias:
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src [tmap:src]
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/src [tmap:src]
Path -> Partition:
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/src
Partition
base file name: src
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name src
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/src
+ name src
serialization.ddl struct src { string key, string value}
serialization.format 1
- columns key,value
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
- transient_lastDdlTime 1262137436
+ transient_lastDdlTime 1266042776
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name src
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/src
+ name src
serialization.ddl struct src { string key, string value}
serialization.format 1
- columns key,value
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
- transient_lastDdlTime 1262137436
+ transient_lastDdlTime 1266042776
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: src
name: src
@@ -123,14 +123,14 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/237484840/10001
+ directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-57_219_3180291446209228569/10001
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
columns _col0,_col1
- serialization.format 1
columns.types string:string
+ serialization.format 1
Stage: Stage-0
Fetch Operator
@@ -146,7 +146,7 @@
SELECT tmap.key, regexp_extract(tmap.value, 'val_(\\d+\\t\\d+)',1) WHERE tmap.key < 100
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/687662347/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-57_269_5995859906691510320/10000
POSTHOOK: query: FROM (
FROM src
SELECT TRANSFORM(src.key, src.value, 1+2, 3+4)
@@ -156,7 +156,7 @@
SELECT tmap.key, regexp_extract(tmap.value, 'val_(\\d+\\t\\d+)',1) WHERE tmap.key < 100
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/687662347/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-57_269_5995859906691510320/10000
0 0 3
0 0 3
0 0 3
@@ -290,10 +290,10 @@
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- field.delim 9
columns _col0,_col1
- serialization.format 9
columns.types string,string
+ field.delim 9
+ serialization.format 9
serialization.last.column.takes.rest true
Reduce Output Operator
key expressions:
@@ -311,41 +311,41 @@
type: string
Needs Tagging: false
Path -> Alias:
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src [tmap:src]
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/src [tmap:src]
Path -> Partition:
- file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
+ file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/src
Partition
base file name: src
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name src
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/src
+ name src
serialization.ddl struct src { string key, string value}
serialization.format 1
- columns key,value
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
- transient_lastDdlTime 1262137436
+ transient_lastDdlTime 1266042776
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
- name src
+ bucket_count -1
+ columns key,value
columns.types string:string
+ file.inputformat org.apache.hadoop.mapred.TextInputFormat
+ file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/src
+ name src
serialization.ddl struct src { string key, string value}
serialization.format 1
- columns key,value
- bucket_count -1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- file.inputformat org.apache.hadoop.mapred.TextInputFormat
- file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
- transient_lastDdlTime 1262137436
+ transient_lastDdlTime 1266042776
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: src
name: src
@@ -366,14 +366,14 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1239765305/10001
+ directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-33-03_106_231065528465301889/10001
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
columns _col0,_col1
- serialization.format 1
columns.types string:string
+ serialization.format 1
Stage: Stage-0
Fetch Operator
@@ -389,7 +389,7 @@
SELECT tmap.key, regexp_extract(tmap.value, 'val_(\\d+\\t\\d+)') WHERE tmap.key < 100
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1514486187/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-33-03_163_5396716588759691796/10000
POSTHOOK: query: FROM (
FROM src
SELECT TRANSFORM(src.key, src.value, 1+2, 3+4)
@@ -399,7 +399,7 @@
SELECT tmap.key, regexp_extract(tmap.value, 'val_(\\d+\\t\\d+)') WHERE tmap.key < 100
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1514486187/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-33-03_163_5396716588759691796/10000
0 0 3
0 0 3
0 0 3