You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by nz...@apache.org on 2010/02/14 03:31:44 UTC

svn commit: r909965 [9/15] - in /hadoop/hive/trunk: ./ contrib/src/test/results/clientnegative/ contrib/src/test/results/clientpositive/ ql/src/java/org/apache/hadoop/hive/ql/exec/ ql/src/test/results/clientpositive/

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_join3.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_join3.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_join3.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_join3.q.out Sun Feb 14 02:31:40 2010
@@ -33,38 +33,6 @@
   Stage: Stage-1
     Map Reduce
       Alias -> Map Operator Tree:
-        src2:src 
-          TableScan
-            alias: src
-            Filter Operator
-              predicate:
-                  expr: ((key <> '12') and (key <> '4'))
-                  type: boolean
-              Filter Operator
-                predicate:
-                    expr: (key <> '12')
-                    type: boolean
-                Select Operator
-                  expressions:
-                        expr: key
-                        type: string
-                        expr: value
-                        type: string
-                  outputColumnNames: _col0, _col1
-                  Reduce Output Operator
-                    key expressions:
-                          expr: _col0
-                          type: string
-                    sort order: +
-                    Map-reduce partition columns:
-                          expr: _col0
-                          type: string
-                    tag: 1
-                    value expressions:
-                          expr: _col0
-                          type: string
-                          expr: _col1
-                          type: string
         src1:src 
           TableScan
             alias: src
@@ -101,6 +69,38 @@
                             type: string
                             expr: _col1
                             type: string
+        src2:src 
+          TableScan
+            alias: src
+            Filter Operator
+              predicate:
+                  expr: ((key <> '12') and (key <> '4'))
+                  type: boolean
+              Filter Operator
+                predicate:
+                    expr: (key <> '12')
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: key
+                        type: string
+                        expr: value
+                        type: string
+                  outputColumnNames: _col0, _col1
+                  Reduce Output Operator
+                    key expressions:
+                          expr: _col0
+                          type: string
+                    sort order: +
+                    Map-reduce partition columns:
+                          expr: _col0
+                          type: string
+                    tag: 1
+                    value expressions:
+                          expr: _col0
+                          type: string
+                          expr: _col1
+                          type: string
         src3:src 
           TableScan
             alias: src
@@ -175,7 +175,7 @@
 WHERE src1.c1 > '0' and (src1.c2 <> 'val_500' or src1.c1 > '1') and (src2.c3 > '10' or src1.c1 <> '10') and (src2.c3 <> '4') and (src3.c5 <> '1')
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/543319193/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_3/build/ql/scratchdir/hive_2010-02-12_22-18-31_804_5890089031375913660/10000
 POSTHOOK: query: SELECT src1.c1, src2.c4 
 FROM
 (SELECT src.key as c1, src.value as c2 from src where src.key <> '11' ) src1
@@ -188,7 +188,7 @@
 WHERE src1.c1 > '0' and (src1.c2 <> 'val_500' or src1.c1 > '1') and (src2.c3 > '10' or src1.c1 <> '10') and (src2.c3 <> '4') and (src3.c5 <> '1')
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/543319193/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_3/build/ql/scratchdir/hive_2010-02-12_22-18-31_804_5890089031375913660/10000
 100	val_100
 100	val_100
 100	val_100

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_multi_insert.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_multi_insert.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_multi_insert.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_multi_insert.q.out Sun Feb 14 02:31:40 2010
@@ -53,9 +53,9 @@
   Stage: Stage-4
     Map Reduce
       Alias -> Map Operator Tree:
-        b 
+        a 
           TableScan
-            alias: b
+            alias: a
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -64,10 +64,15 @@
               Map-reduce partition columns:
                     expr: key
                     type: string
-              tag: 1
-        a 
+              tag: 0
+              value expressions:
+                    expr: key
+                    type: string
+                    expr: value
+                    type: string
+        b 
           TableScan
-            alias: a
+            alias: b
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -76,12 +81,7 @@
               Map-reduce partition columns:
                     expr: key
                     type: string
-              tag: 0
-              value expressions:
-                    expr: key
-                    type: string
-                    expr: value
-                    type: string
+              tag: 1
       Reduce Operator Tree:
         Join Operator
           condition map:
@@ -246,11 +246,11 @@
 PREHOOK: query: SELECT mi1.* FROM mi1
 PREHOOK: type: QUERY
 PREHOOK: Input: default@mi1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/775375392/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-34-41_902_6905591894666460215/10000
 POSTHOOK: query: SELECT mi1.* FROM mi1
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@mi1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/775375392/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-34-41_902_6905591894666460215/10000
 0	val_0
 0	val_0
 0	val_0
@@ -402,11 +402,11 @@
 PREHOOK: query: SELECT mi2.* FROM mi2
 PREHOOK: type: QUERY
 PREHOOK: Input: default@mi2
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/106849464/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-34-41_936_9202557721170644347/10000
 POSTHOOK: query: SELECT mi2.* FROM mi2
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@mi2
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/106849464/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-34-41_936_9202557721170644347/10000
 100	val_100
 100	val_100
 100	val_100
@@ -621,11 +621,11 @@
 PREHOOK: query: SELECT mi3.* FROM mi3
 PREHOOK: type: QUERY
 PREHOOK: Input: default@mi3@ds=2008-04-08/hr=12
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1706731089/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-34-41_968_153693361750093127/10000
 POSTHOOK: query: SELECT mi3.* FROM mi3
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@mi3@ds=2008-04-08/hr=12
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1706731089/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-34-41_968_153693361750093127/10000
 200	2008-04-08	12
 200	2008-04-08	12
 200	2008-04-08	12

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join1.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join1.q.out Sun Feb 14 02:31:40 2010
@@ -27,23 +27,6 @@
   Stage: Stage-1
     Map Reduce
       Alias -> Map Operator Tree:
-        b 
-          TableScan
-            alias: b
-            Reduce Output Operator
-              key expressions:
-                    expr: key
-                    type: string
-              sort order: +
-              Map-reduce partition columns:
-                    expr: key
-                    type: string
-              tag: 1
-              value expressions:
-                    expr: key
-                    type: string
-                    expr: value
-                    type: string
         a 
           TableScan
             alias: a
@@ -65,6 +48,23 @@
                       type: string
                       expr: value
                       type: string
+        b 
+          TableScan
+            alias: b
+            Reduce Output Operator
+              key expressions:
+                    expr: key
+                    type: string
+              sort order: +
+              Map-reduce partition columns:
+                    expr: key
+                    type: string
+              tag: 1
+              value expressions:
+                    expr: key
+                    type: string
+                    expr: value
+                    type: string
       Reduce Operator Tree:
         Join Operator
           condition map:
@@ -110,7 +110,7 @@
  WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/244410137/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_3/build/ql/scratchdir/hive_2010-02-12_22-18-41_861_2988494112154470083/10000
 POSTHOOK: query: FROM 
   src a
  LEFT OUTER JOIN 
@@ -120,7 +120,7 @@
  WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/244410137/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_3/build/ql/scratchdir/hive_2010-02-12_22-18-41_861_2988494112154470083/10000
 17	val_17	17	val_17
 18	val_18	18	val_18
 18	val_18	18	val_18

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join2.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join2.q.out Sun Feb 14 02:31:40 2010
@@ -27,6 +27,23 @@
   Stage: Stage-1
     Map Reduce
       Alias -> Map Operator Tree:
+        a 
+          TableScan
+            alias: a
+            Reduce Output Operator
+              key expressions:
+                    expr: key
+                    type: string
+              sort order: +
+              Map-reduce partition columns:
+                    expr: key
+                    type: string
+              tag: 0
+              value expressions:
+                    expr: key
+                    type: string
+                    expr: value
+                    type: string
         b 
           TableScan
             alias: b
@@ -48,23 +65,6 @@
                       type: string
                       expr: value
                       type: string
-        a 
-          TableScan
-            alias: a
-            Reduce Output Operator
-              key expressions:
-                    expr: key
-                    type: string
-              sort order: +
-              Map-reduce partition columns:
-                    expr: key
-                    type: string
-              tag: 0
-              value expressions:
-                    expr: key
-                    type: string
-                    expr: value
-                    type: string
       Reduce Operator Tree:
         Join Operator
           condition map:
@@ -110,7 +110,7 @@
  WHERE a.key > '10' AND a.key < '20' AND b.key > '15' AND b.key < '25'
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/1607355712/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_3/build/ql/scratchdir/hive_2010-02-12_22-18-49_166_7655816003584089890/10000
 POSTHOOK: query: FROM 
   src a
  RIGHT OUTER JOIN 
@@ -120,7 +120,7 @@
  WHERE a.key > '10' AND a.key < '20' AND b.key > '15' AND b.key < '25'
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/1607355712/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_3/build/ql/scratchdir/hive_2010-02-12_22-18-49_166_7655816003584089890/10000
 150	val_150	150	val_150
 152	val_152	152	val_152
 152	val_152	152	val_152

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join3.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join3.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join3.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join3.q.out Sun Feb 14 02:31:40 2010
@@ -27,9 +27,9 @@
   Stage: Stage-1
     Map Reduce
       Alias -> Map Operator Tree:
-        b 
+        a 
           TableScan
-            alias: b
+            alias: a
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -38,15 +38,15 @@
               Map-reduce partition columns:
                     expr: key
                     type: string
-              tag: 1
+              tag: 0
               value expressions:
                     expr: key
                     type: string
                     expr: value
                     type: string
-        a 
+        b 
           TableScan
-            alias: a
+            alias: b
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -55,7 +55,7 @@
               Map-reduce partition columns:
                     expr: key
                     type: string
-              tag: 0
+              tag: 1
               value expressions:
                     expr: key
                     type: string
@@ -106,7 +106,7 @@
  WHERE a.key > '10' AND a.key < '20' AND b.key > '15' AND b.key < '25'
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/794158250/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_3/build/ql/scratchdir/hive_2010-02-12_22-18-58_300_1623871517298833290/10000
 POSTHOOK: query: FROM 
   src a
  FULL OUTER JOIN 
@@ -116,7 +116,7 @@
  WHERE a.key > '10' AND a.key < '20' AND b.key > '15' AND b.key < '25'
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/794158250/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_3/build/ql/scratchdir/hive_2010-02-12_22-18-58_300_1623871517298833290/10000
 150	val_150	150	val_150
 152	val_152	152	val_152
 152	val_152	152	val_152

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join4.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join4.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join4.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join4.q.out Sun Feb 14 02:31:40 2010
@@ -33,9 +33,9 @@
   Stage: Stage-1
     Map Reduce
       Alias -> Map Operator Tree:
-        b 
+        a 
           TableScan
-            alias: b
+            alias: a
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -44,15 +44,15 @@
               Map-reduce partition columns:
                     expr: key
                     type: string
-              tag: 1
+              tag: 0
               value expressions:
                     expr: key
                     type: string
                     expr: value
                     type: string
-        c 
+        b 
           TableScan
-            alias: c
+            alias: b
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -61,13 +61,15 @@
               Map-reduce partition columns:
                     expr: key
                     type: string
-              tag: 2
+              tag: 1
               value expressions:
                     expr: key
                     type: string
-        a 
+                    expr: value
+                    type: string
+        c 
           TableScan
-            alias: a
+            alias: c
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -76,12 +78,10 @@
               Map-reduce partition columns:
                     expr: key
                     type: string
-              tag: 0
+              tag: 2
               value expressions:
                     expr: key
                     type: string
-                    expr: value
-                    type: string
       Reduce Operator Tree:
         Join Operator
           condition map:
@@ -134,7 +134,7 @@
  WHERE a.key > '10' AND a.key < '20' AND b.key > '15' AND b.key < '25' AND sqrt(c.key) <> 13
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/1320647695/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-31-48_596_2953182954930521327/10000
 POSTHOOK: query: FROM 
   src a
  LEFT OUTER JOIN
@@ -147,7 +147,7 @@
  WHERE a.key > '10' AND a.key < '20' AND b.key > '15' AND b.key < '25' AND sqrt(c.key) <> 13
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/1320647695/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-31-48_596_2953182954930521327/10000
 150	val_150	150	val_150	150
 152	val_152	152	val_152	152
 152	val_152	152	val_152	152

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_random.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_random.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_random.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_random.q.out Sun Feb 14 02:31:40 2010
@@ -27,6 +27,26 @@
   Stage: Stage-1
     Map Reduce
       Alias -> Map Operator Tree:
+        src1:src 
+          TableScan
+            alias: src
+            Select Operator
+              expressions:
+                    expr: key
+                    type: string
+              outputColumnNames: _col0
+              Reduce Output Operator
+                key expressions:
+                      expr: _col0
+                      type: string
+                sort order: +
+                Map-reduce partition columns:
+                      expr: _col0
+                      type: string
+                tag: 0
+                value expressions:
+                      expr: _col0
+                      type: string
         src2:src 
           TableScan
             alias: src
@@ -57,26 +77,6 @@
                     value expressions:
                           expr: _col1
                           type: string
-        src1:src 
-          TableScan
-            alias: src
-            Select Operator
-              expressions:
-                    expr: key
-                    type: string
-              outputColumnNames: _col0
-              Reduce Output Operator
-                key expressions:
-                      expr: _col0
-                      type: string
-                sort order: +
-                Map-reduce partition columns:
-                      expr: _col0
-                      type: string
-                tag: 0
-                value expressions:
-                      expr: _col0
-                      type: string
       Reduce Operator Tree:
         Join Operator
           condition map:

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_udf_case.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_udf_case.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_udf_case.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_udf_case.q.out Sun Feb 14 02:31:40 2010
@@ -36,12 +36,12 @@
   Stage: Stage-1
     Map Reduce
       Alias -> Map Operator Tree:
-        b 
+        a 
           TableScan
-            alias: b
+            alias: a
             Filter Operator
               predicate:
-                  expr: (ds = '2008-04-08')
+                  expr: ((ds = '2008-04-08') and CASE (key) WHEN ('27') THEN (true) WHEN ('38') THEN (false) ELSE (null) END)
                   type: boolean
               Reduce Output Operator
                 key expressions:
@@ -51,7 +51,7 @@
                 Map-reduce partition columns:
                       expr: key
                       type: string
-                tag: 1
+                tag: 0
                 value expressions:
                       expr: key
                       type: string
@@ -61,12 +61,12 @@
                       type: string
                       expr: hr
                       type: string
-        a 
+        b 
           TableScan
-            alias: a
+            alias: b
             Filter Operator
               predicate:
-                  expr: ((ds = '2008-04-08') and CASE (key) WHEN ('27') THEN (true) WHEN ('38') THEN (false) ELSE (null) END)
+                  expr: (ds = '2008-04-08')
                   type: boolean
               Reduce Output Operator
                 key expressions:
@@ -76,7 +76,7 @@
                 Map-reduce partition columns:
                       expr: key
                       type: string
-                tag: 0
+                tag: 1
                 value expressions:
                       expr: key
                       type: string
@@ -128,7 +128,7 @@
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/438623182/10002 
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-31-56_689_3254720341447490427/10002 
             Reduce Output Operator
               key expressions:
                     expr: _col0
@@ -194,7 +194,7 @@
 PREHOOK: type: QUERY
 PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
 PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/2137110555/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-31-56_869_4629499201500113908/10000
 POSTHOOK: query: SELECT *
 FROM srcpart a JOIN srcpart b
 ON a.key = b.key
@@ -209,7 +209,7 @@
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
 POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/2137110555/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-31-56_869_4629499201500113908/10000
 27	val_27	2008-04-08	11	27	val_27	2008-04-08	11
 27	val_27	2008-04-08	11	27	val_27	2008-04-08	12
 27	val_27	2008-04-08	12	27	val_27	2008-04-08	11

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner1.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner1.q.out Sun Feb 14 02:31:40 2010
@@ -33,51 +33,51 @@
                 File Output Operator
                   compressed: false
                   GlobalTableId: 0
-                  directory: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1281920331/10001
+                  directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-35-08_059_3516688999660848477/10001
                   table:
                       input format: org.apache.hadoop.mapred.TextInputFormat
                       output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                       properties:
                         columns _col0,_col1
-                        serialization.format 1
                         columns.types string:string
+                        serialization.format 1
       Needs Tagging: false
       Path -> Alias:
-        file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src [src]
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/src [src]
       Path -> Partition:
-        file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src 
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/src 
           Partition
             base file name: src
             input format: org.apache.hadoop.mapred.TextInputFormat
             output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
             properties:
-              name src
+              bucket_count -1
+              columns key,value
               columns.types string:string
+              file.inputformat org.apache.hadoop.mapred.TextInputFormat
+              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/src
+              name src
               serialization.ddl struct src { string key, string value}
               serialization.format 1
-              columns key,value
-              bucket_count -1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-              file.inputformat org.apache.hadoop.mapred.TextInputFormat
-              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-              location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
-              transient_lastDdlTime 1262137321
+              transient_lastDdlTime 1266042907
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
               properties:
-                name src
+                bucket_count -1
+                columns key,value
                 columns.types string:string
+                file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/test/data/warehouse/src
+                name src
                 serialization.ddl struct src { string key, string value}
                 serialization.format 1
-                columns key,value
-                bucket_count -1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
-                transient_lastDdlTime 1262137321
+                transient_lastDdlTime 1266042907
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: src
             name: src
@@ -90,11 +90,11 @@
 PREHOOK: query: select * from src where rand(1) < 0.1
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1343937738/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-35-08_099_7501446808782656360/10000
 POSTHOOK: query: select * from src where rand(1) < 0.1
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1343937738/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-35-08_099_7501446808782656360/10000
 409	val_409
 429	val_429
 209	val_209

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner2.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner2.q.out Sun Feb 14 02:31:40 2010
@@ -53,30 +53,30 @@
                 File Output Operator
                   compressed: false
                   GlobalTableId: 1
-                  directory: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/297465583/10002
+                  directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-21_884_2937772000010325968/10002
                   table:
                       input format: org.apache.hadoop.mapred.TextInputFormat
                       output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                       properties:
-                        name tmptable
+                        bucket_count -1
+                        columns key,value,hr,ds
                         columns.types string:string:string:string
+                        file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                        file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/tmptable
+                        name tmptable
                         serialization.ddl struct tmptable { string key, string value, string hr, string ds}
                         serialization.format 1
-                        columns key,value,hr,ds
-                        bucket_count -1
                         serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                        file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                        file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                        location file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/tmptable
-                        transient_lastDdlTime 1263595477
+                        transient_lastDdlTime 1266042741
                       serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                       name: tmptable
       Needs Tagging: false
       Path -> Alias:
-        file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [a]
-        file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [a]
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [a]
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [a]
       Path -> Partition:
-        file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 
           Partition
             base file name: hr=11
             input format: org.apache.hadoop.mapred.TextInputFormat
@@ -85,39 +85,39 @@
               ds 2008-04-08
               hr 11
             properties:
-              name srcpart
+              bucket_count -1
+              columns key,value
               columns.types string:string
+              file.inputformat org.apache.hadoop.mapred.TextInputFormat
+              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/srcpart
+              name srcpart
+              partition_columns ds/hr
               serialization.ddl struct srcpart { string key, string value}
               serialization.format 1
-              columns key,value
-              partition_columns ds/hr
-              bucket_count -1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-              file.inputformat org.apache.hadoop.mapred.TextInputFormat
-              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-              location file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/srcpart
-              transient_lastDdlTime 1263595476
+              transient_lastDdlTime 1266042740
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
               properties:
-                name srcpart
+                bucket_count -1
+                columns key,value
                 columns.types string:string
+                file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/srcpart
+                name srcpart
+                partition_columns ds/hr
                 serialization.ddl struct srcpart { string key, string value}
                 serialization.format 1
-                columns key,value
-                partition_columns ds/hr
-                bucket_count -1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/srcpart
-                transient_lastDdlTime 1263595476
+                transient_lastDdlTime 1266042740
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: srcpart
             name: srcpart
-        file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 
           Partition
             base file name: hr=12
             input format: org.apache.hadoop.mapred.TextInputFormat
@@ -126,35 +126,35 @@
               ds 2008-04-08
               hr 12
             properties:
-              name srcpart
+              bucket_count -1
+              columns key,value
               columns.types string:string
+              file.inputformat org.apache.hadoop.mapred.TextInputFormat
+              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/srcpart
+              name srcpart
+              partition_columns ds/hr
               serialization.ddl struct srcpart { string key, string value}
               serialization.format 1
-              columns key,value
-              partition_columns ds/hr
-              bucket_count -1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-              file.inputformat org.apache.hadoop.mapred.TextInputFormat
-              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-              location file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/srcpart
-              transient_lastDdlTime 1263595476
+              transient_lastDdlTime 1266042740
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
               properties:
-                name srcpart
+                bucket_count -1
+                columns key,value
                 columns.types string:string
+                file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/srcpart
+                name srcpart
+                partition_columns ds/hr
                 serialization.ddl struct srcpart { string key, string value}
                 serialization.format 1
-                columns key,value
-                partition_columns ds/hr
-                bucket_count -1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/srcpart
-                transient_lastDdlTime 1263595476
+                transient_lastDdlTime 1266042740
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: srcpart
             name: srcpart
@@ -166,37 +166,37 @@
     Move Operator
       files:
           hdfs directory: true
-          source: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/297465583/10002
-          destination: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/567762797/10000
+          source: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-21_884_2937772000010325968/10002
+          destination: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-21_884_2937772000010325968/10000
 
   Stage: Stage-0
     Move Operator
       tables:
           replace: true
-          source: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/567762797/10000
+          source: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-21_884_2937772000010325968/10000
           table:
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
               properties:
-                name tmptable
+                bucket_count -1
+                columns key,value,hr,ds
                 columns.types string:string:string:string
+                file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/tmptable
+                name tmptable
                 serialization.ddl struct tmptable { string key, string value, string hr, string ds}
                 serialization.format 1
-                columns key,value,hr,ds
-                bucket_count -1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/tmptable
-                transient_lastDdlTime 1263595477
+                transient_lastDdlTime 1266042741
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: tmptable
-          tmp directory: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/567762797/10001
+          tmp directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-21_884_2937772000010325968/10001
 
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/297465583/10002 
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-21_884_2937772000010325968/10002 
             Reduce Output Operator
               sort order: 
               Map-reduce partition columns:
@@ -214,41 +214,41 @@
                     type: string
       Needs Tagging: false
       Path -> Alias:
-        file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/297465583/10002 [file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/297465583/10002]
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-21_884_2937772000010325968/10002 [file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-21_884_2937772000010325968/10002]
       Path -> Partition:
-        file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/297465583/10002 
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-21_884_2937772000010325968/10002 
           Partition
             base file name: 10002
             input format: org.apache.hadoop.mapred.TextInputFormat
             output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
             properties:
-              name tmptable
+              bucket_count -1
+              columns key,value,hr,ds
               columns.types string:string:string:string
+              file.inputformat org.apache.hadoop.mapred.TextInputFormat
+              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/tmptable
+              name tmptable
               serialization.ddl struct tmptable { string key, string value, string hr, string ds}
               serialization.format 1
-              columns key,value,hr,ds
-              bucket_count -1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-              file.inputformat org.apache.hadoop.mapred.TextInputFormat
-              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-              location file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/tmptable
-              transient_lastDdlTime 1263595477
+              transient_lastDdlTime 1266042741
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
               properties:
-                name tmptable
+                bucket_count -1
+                columns key,value,hr,ds
                 columns.types string:string:string:string
+                file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/tmptable
+                name tmptable
                 serialization.ddl struct tmptable { string key, string value, string hr, string ds}
                 serialization.format 1
-                columns key,value,hr,ds
-                bucket_count -1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/tmptable
-                transient_lastDdlTime 1263595477
+                transient_lastDdlTime 1266042741
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: tmptable
             name: tmptable
@@ -257,22 +257,22 @@
           File Output Operator
             compressed: false
             GlobalTableId: 0
-            directory: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/567762797/10000
+            directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-21_884_2937772000010325968/10000
             table:
                 input format: org.apache.hadoop.mapred.TextInputFormat
                 output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                 properties:
-                  name tmptable
+                  bucket_count -1
+                  columns key,value,hr,ds
                   columns.types string:string:string:string
+                  file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                  file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                  location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/tmptable
+                  name tmptable
                   serialization.ddl struct tmptable { string key, string value, string hr, string ds}
                   serialization.format 1
-                  columns key,value,hr,ds
-                  bucket_count -1
                   serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                  file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                  location file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/test/data/warehouse/tmptable
-                  file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  transient_lastDdlTime 1263595477
+                  transient_lastDdlTime 1266042741
                 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                 name: tmptable
 
@@ -292,11 +292,11 @@
 PREHOOK: query: select * from tmptable x sort by x.key,x.value,x.ds,x.hr
 PREHOOK: type: QUERY
 PREHOOK: Input: default@tmptable
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/543578106/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-32_891_7737984582274649304/10000
 POSTHOOK: query: select * from tmptable x sort by x.key,x.value,x.ds,x.hr
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@tmptable
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/543578106/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-32_891_7737984582274649304/10000
 103	val_103	2008-04-08	11
 103	val_103	2008-04-08	12
 133	val_133	2008-04-08	11

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner3.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner3.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner3.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner3.q.out Sun Feb 14 02:31:40 2010
@@ -39,19 +39,19 @@
                 File Output Operator
                   compressed: false
                   GlobalTableId: 0
-                  directory: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/124088894/10001
+                  directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-37-59_832_3788417974183008166/10001
                   table:
                       input format: org.apache.hadoop.mapred.TextInputFormat
                       output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                       properties:
                         columns _col0,_col1,_col2,_col3
-                        serialization.format 1
                         columns.types string:string:string:string
+                        serialization.format 1
       Needs Tagging: false
       Path -> Alias:
-        file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [a]
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [a]
       Path -> Partition:
-        file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 
           Partition
             base file name: hr=12
             input format: org.apache.hadoop.mapred.TextInputFormat
@@ -60,35 +60,35 @@
               ds 2008-04-08
               hr 12
             properties:
-              name srcpart
+              bucket_count -1
+              columns key,value
               columns.types string:string
+              file.inputformat org.apache.hadoop.mapred.TextInputFormat
+              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart
+              name srcpart
+              partition_columns ds/hr
               serialization.ddl struct srcpart { string key, string value}
               serialization.format 1
-              columns key,value
-              partition_columns ds/hr
-              bucket_count -1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-              file.inputformat org.apache.hadoop.mapred.TextInputFormat
-              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-              location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart
-              transient_lastDdlTime 1262137336
+              transient_lastDdlTime 1266043078
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
               properties:
-                name srcpart
+                bucket_count -1
+                columns key,value
                 columns.types string:string
+                file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart
+                name srcpart
+                partition_columns ds/hr
                 serialization.ddl struct srcpart { string key, string value}
                 serialization.format 1
-                columns key,value
-                partition_columns ds/hr
-                bucket_count -1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart
-                transient_lastDdlTime 1262137336
+                transient_lastDdlTime 1266043078
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: srcpart
             name: srcpart
@@ -101,11 +101,11 @@
 PREHOOK: query: select a.* from srcpart a where rand(1) < 0.1 and a.ds = '2008-04-08' and not(key > 50 or key < 10) and a.hr like '%2'
 PREHOOK: type: QUERY
 PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-PREHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1944369228/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-37-59_919_2841312693651221074/10000
 POSTHOOK: query: select a.* from srcpart a where rand(1) < 0.1 and a.ds = '2008-04-08' and not(key > 50 or key < 10) and a.hr like '%2'
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-POSTHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1944369228/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-37-59_919_2841312693651221074/10000
 42	val_42	2008-04-08	12
 44	val_44	2008-04-08	12
 26	val_26	2008-04-08	12
@@ -155,19 +155,19 @@
                   File Output Operator
                     compressed: false
                     GlobalTableId: 0
-                    directory: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1956461194/10001
+                    directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-38-03_773_7561053502233744490/10001
                     table:
                         input format: org.apache.hadoop.mapred.TextInputFormat
                         output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                         properties:
                           columns _col0,_col1,_col2,_col3
-                          serialization.format 1
                           columns.types string:string:string:string
+                          serialization.format 1
       Needs Tagging: false
       Path -> Alias:
-        file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [a]
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [a]
       Path -> Partition:
-        file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 
           Partition
             base file name: hr=12
             input format: org.apache.hadoop.mapred.TextInputFormat
@@ -176,35 +176,35 @@
               ds 2008-04-08
               hr 12
             properties:
-              name srcpart
+              bucket_count -1
+              columns key,value
               columns.types string:string
+              file.inputformat org.apache.hadoop.mapred.TextInputFormat
+              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart
+              name srcpart
+              partition_columns ds/hr
               serialization.ddl struct srcpart { string key, string value}
               serialization.format 1
-              columns key,value
-              partition_columns ds/hr
-              bucket_count -1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-              file.inputformat org.apache.hadoop.mapred.TextInputFormat
-              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-              location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart
-              transient_lastDdlTime 1262137336
+              transient_lastDdlTime 1266043078
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
               properties:
-                name srcpart
+                bucket_count -1
+                columns key,value
                 columns.types string:string
+                file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart
+                name srcpart
+                partition_columns ds/hr
                 serialization.ddl struct srcpart { string key, string value}
                 serialization.format 1
-                columns key,value
-                partition_columns ds/hr
-                bucket_count -1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/srcpart
-                transient_lastDdlTime 1262137336
+                transient_lastDdlTime 1266043078
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: srcpart
             name: srcpart
@@ -217,11 +217,11 @@
 PREHOOK: query: select a.* from srcpart a where a.ds = '2008-04-08' and not(key > 50 or key < 10) and a.hr like '%2'
 PREHOOK: type: QUERY
 PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-PREHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1681163814/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-38-03_865_149659006171751209/10000
 POSTHOOK: query: select a.* from srcpart a where a.ds = '2008-04-08' and not(key > 50 or key < 10) and a.hr like '%2'
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-POSTHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1681163814/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-38-03_865_149659006171751209/10000
 27	val_27	2008-04-08	12
 37	val_37	2008-04-08	12
 15	val_15	2008-04-08	12

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/regex_col.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/regex_col.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/regex_col.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/regex_col.q.out Sun Feb 14 02:31:40 2010
@@ -113,9 +113,9 @@
   Stage: Stage-1
     Map Reduce
       Alias -> Map Operator Tree:
-        b 
+        a 
           TableScan
-            alias: b
+            alias: a
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -128,15 +128,15 @@
                     type: string
                     expr: value
                     type: string
-              tag: 1
+              tag: 0
               value expressions:
                     expr: ds
                     type: string
                     expr: hr
                     type: string
-        a 
+        b 
           TableScan
-            alias: a
+            alias: b
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -149,7 +149,7 @@
                     type: string
                     expr: value
                     type: string
-              tag: 0
+              tag: 1
               value expressions:
                     expr: ds
                     type: string
@@ -209,31 +209,6 @@
   Stage: Stage-1
     Map Reduce
       Alias -> Map Operator Tree:
-        b 
-          TableScan
-            alias: b
-            Reduce Output Operator
-              key expressions:
-                    expr: key
-                    type: string
-                    expr: hr
-                    type: string
-                    expr: ds
-                    type: string
-              sort order: +++
-              Map-reduce partition columns:
-                    expr: key
-                    type: string
-                    expr: hr
-                    type: string
-                    expr: ds
-                    type: string
-              tag: 1
-              value expressions:
-                    expr: ds
-                    type: string
-                    expr: hr
-                    type: string
         a 
           TableScan
             alias: a
@@ -262,6 +237,31 @@
                         expr: ds
                         type: string
                   tag: 0
+        b 
+          TableScan
+            alias: b
+            Reduce Output Operator
+              key expressions:
+                    expr: key
+                    type: string
+                    expr: hr
+                    type: string
+                    expr: ds
+                    type: string
+              sort order: +++
+              Map-reduce partition columns:
+                    expr: key
+                    type: string
+                    expr: hr
+                    type: string
+                    expr: ds
+                    type: string
+              tag: 1
+              value expressions:
+                    expr: ds
+                    type: string
+                    expr: hr
+                    type: string
       Reduce Operator Tree:
         Join Operator
           condition map:
@@ -288,7 +288,7 @@
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/832608785/10002 
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-36-24_399_2804249481999873947/10002 
             Reduce Output Operator
               key expressions:
                     expr: _col0
@@ -324,7 +324,7 @@
 PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
 PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
 PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1495752848/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-36-24_582_4644469832031285008/10000
 POSTHOOK: query: SELECT b.`..` FROM srcpart a JOIN srcpart b
 ON a.key = b.key AND a.hr = b.hr AND a.ds = b.ds AND a.key = 103
 ORDER BY ds, hr
@@ -333,7 +333,7 @@
 POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
 POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
 POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1495752848/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-36-24_582_4644469832031285008/10000
 2008-04-08	11
 2008-04-08	11
 2008-04-08	11
@@ -526,14 +526,14 @@
 PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
 PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
 PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/316243993/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-36-36_271_5109114395556786366/10000
 POSTHOOK: query: SELECT `(ds|hr)?+.+` FROM srcpart ORDER BY key, value LIMIT 10
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
 POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
 POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
 POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/316243993/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-36-36_271_5109114395556786366/10000
 0	val_0
 0	val_0
 0	val_0

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/regexp_extract.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/regexp_extract.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/regexp_extract.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/regexp_extract.q.out Sun Feb 14 02:31:40 2010
@@ -47,10 +47,10 @@
                     input format: org.apache.hadoop.mapred.TextInputFormat
                     output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                     properties:
-                      field.delim 9
                       columns _col0,_col1
-                      serialization.format 9
                       columns.types string,string
+                      field.delim 9
+                      serialization.format 9
                       serialization.last.column.takes.rest true
                 Reduce Output Operator
                   key expressions:
@@ -68,41 +68,41 @@
                         type: string
       Needs Tagging: false
       Path -> Alias:
-        file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src [tmap:src]
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/src [tmap:src]
       Path -> Partition:
-        file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src 
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/src 
           Partition
             base file name: src
             input format: org.apache.hadoop.mapred.TextInputFormat
             output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
             properties:
-              name src
+              bucket_count -1
+              columns key,value
               columns.types string:string
+              file.inputformat org.apache.hadoop.mapred.TextInputFormat
+              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/src
+              name src
               serialization.ddl struct src { string key, string value}
               serialization.format 1
-              columns key,value
-              bucket_count -1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-              file.inputformat org.apache.hadoop.mapred.TextInputFormat
-              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-              location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
-              transient_lastDdlTime 1262137436
+              transient_lastDdlTime 1266042776
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
               properties:
-                name src
+                bucket_count -1
+                columns key,value
                 columns.types string:string
+                file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/src
+                name src
                 serialization.ddl struct src { string key, string value}
                 serialization.format 1
-                columns key,value
-                bucket_count -1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
-                transient_lastDdlTime 1262137436
+                transient_lastDdlTime 1266042776
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: src
             name: src
@@ -123,14 +123,14 @@
               File Output Operator
                 compressed: false
                 GlobalTableId: 0
-                directory: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/237484840/10001
+                directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-57_219_3180291446209228569/10001
                 table:
                     input format: org.apache.hadoop.mapred.TextInputFormat
                     output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                     properties:
                       columns _col0,_col1
-                      serialization.format 1
                       columns.types string:string
+                      serialization.format 1
 
   Stage: Stage-0
     Fetch Operator
@@ -146,7 +146,7 @@
 SELECT tmap.key, regexp_extract(tmap.value, 'val_(\\d+\\t\\d+)',1) WHERE tmap.key < 100
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/687662347/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-57_269_5995859906691510320/10000
 POSTHOOK: query: FROM (
   FROM src
   SELECT TRANSFORM(src.key, src.value, 1+2, 3+4)
@@ -156,7 +156,7 @@
 SELECT tmap.key, regexp_extract(tmap.value, 'val_(\\d+\\t\\d+)',1) WHERE tmap.key < 100
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/687662347/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-32-57_269_5995859906691510320/10000
 0	0	3
 0	0	3
 0	0	3
@@ -290,10 +290,10 @@
                     input format: org.apache.hadoop.mapred.TextInputFormat
                     output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                     properties:
-                      field.delim 9
                       columns _col0,_col1
-                      serialization.format 9
                       columns.types string,string
+                      field.delim 9
+                      serialization.format 9
                       serialization.last.column.takes.rest true
                 Reduce Output Operator
                   key expressions:
@@ -311,41 +311,41 @@
                         type: string
       Needs Tagging: false
       Path -> Alias:
-        file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src [tmap:src]
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/src [tmap:src]
       Path -> Partition:
-        file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src 
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/src 
           Partition
             base file name: src
             input format: org.apache.hadoop.mapred.TextInputFormat
             output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
             properties:
-              name src
+              bucket_count -1
+              columns key,value
               columns.types string:string
+              file.inputformat org.apache.hadoop.mapred.TextInputFormat
+              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/src
+              name src
               serialization.ddl struct src { string key, string value}
               serialization.format 1
-              columns key,value
-              bucket_count -1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-              file.inputformat org.apache.hadoop.mapred.TextInputFormat
-              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-              location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
-              transient_lastDdlTime 1262137436
+              transient_lastDdlTime 1266042776
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
               properties:
-                name src
+                bucket_count -1
+                columns key,value
                 columns.types string:string
+                file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/test/data/warehouse/src
+                name src
                 serialization.ddl struct src { string key, string value}
                 serialization.format 1
-                columns key,value
-                bucket_count -1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/jsichi/open/hive-trunk/build/ql/test/data/warehouse/src
-                transient_lastDdlTime 1262137436
+                transient_lastDdlTime 1266042776
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: src
             name: src
@@ -366,14 +366,14 @@
               File Output Operator
                 compressed: false
                 GlobalTableId: 0
-                directory: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1239765305/10001
+                directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-33-03_106_231065528465301889/10001
                 table:
                     input format: org.apache.hadoop.mapred.TextInputFormat
                     output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                     properties:
                       columns _col0,_col1
-                      serialization.format 1
                       columns.types string:string
+                      serialization.format 1
 
   Stage: Stage-0
     Fetch Operator
@@ -389,7 +389,7 @@
 SELECT tmap.key, regexp_extract(tmap.value, 'val_(\\d+\\t\\d+)') WHERE tmap.key < 100
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1514486187/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-33-03_163_5396716588759691796/10000
 POSTHOOK: query: FROM (
   FROM src
   SELECT TRANSFORM(src.key, src.value, 1+2, 3+4)
@@ -399,7 +399,7 @@
 SELECT tmap.key, regexp_extract(tmap.value, 'val_(\\d+\\t\\d+)') WHERE tmap.key < 100
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/jsichi/open/hive-trunk/build/ql/tmp/1514486187/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-33-03_163_5396716588759691796/10000
 0	0	3
 0	0	3
 0	0	3