You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by zs...@apache.org on 2009/08/24 20:23:22 UTC
svn commit: r807330 [3/3] - in /hadoop/hive/trunk: ./
common/src/java/org/apache/hadoop/hive/conf/ conf/
contrib/src/java/org/apache/hadoop/hive/contrib/serde2/
contrib/src/test/queries/clientpositive/
contrib/src/test/results/clientpositive/ eclipse-t...
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce3.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce3.q.out?rev=807330&r1=807329&r2=807330&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce3.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce3.q.out Mon Aug 24 18:23:19 2009
@@ -6,7 +6,7 @@
USING '/bin/cat' AS (tkey, ten, one, tvalue)
SORT BY tvalue, tkey
ABSTRACT SYNTAX TREE:
- (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_TAB dest1)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL src) key) (TOK_FUNCTION TOK_INT (/ (. (TOK_TABLE_OR_COL src) key) 10)) (TOK_FUNCTION TOK_INT (% (. (TOK_TABLE_OR_COL src) key) 10)) (. (TOK_TABLE_OR_COL src) value)) '/bin/cat' (TOK_ALIASLIST tkey ten one tvalue)))) (TOK_SORTBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL tvalue)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL tkey)))))
+ (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_TAB dest1)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL src) key) (TOK_FUNCTION TOK_INT (/ (. (TOK_TABLE_OR_COL src) key) 10)) (TOK_FUNCTION TOK_INT (% (. (TOK_TABLE_OR_COL src) key) 10)) (. (TOK_TABLE_OR_COL src) value)) TOK_SERDE '/bin/cat' TOK_SERDE TOK_RECORDREADER (TOK_ALIASLIST tkey ten one tvalue)))) (TOK_SORTBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL tvalue)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL tkey)))))
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -94,7 +94,7 @@
Output: default/dest1
query: SELECT dest1.* FROM dest1
Input: default/dest1
-Output: file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_0/build/ql/tmp/1933750790/10000
+Output: file:/data/users/njain/hive4/hive4/build/ql/tmp/568476430/10000
0 0 0 val_0
0 0 0 val_0
0 0 0 val_0
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce4.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce4.q.out?rev=807330&r1=807329&r2=807330&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce4.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce4.q.out Mon Aug 24 18:23:19 2009
@@ -7,7 +7,7 @@
DISTRIBUTE BY tvalue, tkey
SORT BY ten DESC, one ASC
ABSTRACT SYNTAX TREE:
- (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_TAB dest1)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL src) key) (TOK_FUNCTION TOK_INT (/ (. (TOK_TABLE_OR_COL src) key) 10)) (TOK_FUNCTION TOK_INT (% (. (TOK_TABLE_OR_COL src) key) 10)) (. (TOK_TABLE_OR_COL src) value)) '/bin/cat' (TOK_ALIASLIST tkey ten one tvalue)))) (TOK_DISTRIBUTEBY (TOK_TABLE_OR_COL tvalue) (TOK_TABLE_OR_COL tkey)) (TOK_SORTBY (TOK_TABSORTCOLNAMEDESC (TOK_TABLE_OR_COL ten)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL one)))))
+ (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_TAB dest1)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL src) key) (TOK_FUNCTION TOK_INT (/ (. (TOK_TABLE_OR_COL src) key) 10)) (TOK_FUNCTION TOK_INT (% (. (TOK_TABLE_OR_COL src) key) 10)) (. (TOK_TABLE_OR_COL src) value)) TOK_SERDE '/bin/cat' TOK_SERDE TOK_RECORDREADER (TOK_ALIASLIST tkey ten one tvalue)))) (TOK_DISTRIBUTEBY (TOK_TABLE_OR_COL tvalue) (TOK_TABLE_OR_COL tkey)) (TOK_SORTBY (TOK_TABSORTCOLNAMEDESC (TOK_TABLE_OR_COL ten)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL one)))))
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -101,7 +101,7 @@
Output: default/dest1
query: SELECT dest1.* FROM dest1
Input: default/dest1
-Output: file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_1/build/ql/tmp/34806120/10000
+Output: file:/data/users/njain/hive4/hive4/build/ql/tmp/1394907068/10000
90 9 0 val_90
90 9 0 val_90
90 9 0 val_90
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce7.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce7.q.out?rev=807330&r1=807329&r2=807330&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce7.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce7.q.out Mon Aug 24 18:23:19 2009
@@ -6,7 +6,7 @@
USING '/bin/cat' AS (k, v, tkey, ten, one, tvalue)
SORT BY tvalue, tkey
ABSTRACT SYNTAX TREE:
- (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_TAB dest1)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (TOK_ALLCOLREF src) (. (TOK_TABLE_OR_COL src) key) (TOK_FUNCTION TOK_INT (/ (. (TOK_TABLE_OR_COL src) key) 10)) (TOK_FUNCTION TOK_INT (% (. (TOK_TABLE_OR_COL src) key) 10)) (. (TOK_TABLE_OR_COL src) value)) '/bin/cat' (TOK_ALIASLIST k v tkey ten one tvalue)))) (TOK_SORTBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL tvalue)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL tkey)))))
+ (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_TAB dest1)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (TOK_ALLCOLREF src) (. (TOK_TABLE_OR_COL src) key) (TOK_FUNCTION TOK_INT (/ (. (TOK_TABLE_OR_COL src) key) 10)) (TOK_FUNCTION TOK_INT (% (. (TOK_TABLE_OR_COL src) key) 10)) (. (TOK_TABLE_OR_COL src) value)) TOK_SERDE '/bin/cat' TOK_SERDE TOK_RECORDREADER (TOK_ALIASLIST k v tkey ten one tvalue)))) (TOK_SORTBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL tvalue)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL tkey)))))
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -106,7 +106,7 @@
Output: default/dest1
query: SELECT dest1.* FROM dest1
Input: default/dest1
-Output: file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_1/build/ql/tmp/1267286735/10000
+Output: file:/data/users/njain/hive4/hive4/build/ql/tmp/1083839683/10000
0 val_0 0 0 0 val_0
0 val_0 0 0 0 val_0
0 val_0 0 0 0 val_0
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce8.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce8.q.out?rev=807330&r1=807329&r2=807330&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce8.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce8.q.out Mon Aug 24 18:23:19 2009
@@ -7,7 +7,7 @@
DISTRIBUTE BY rand(3)
SORT BY tvalue, tkey
ABSTRACT SYNTAX TREE:
- (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_TAB dest1)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (TOK_ALLCOLREF src) (. (TOK_TABLE_OR_COL src) key) (TOK_FUNCTION TOK_INT (/ (. (TOK_TABLE_OR_COL src) key) 10)) (TOK_FUNCTION TOK_INT (% (. (TOK_TABLE_OR_COL src) key) 10)) (. (TOK_TABLE_OR_COL src) value)) '/bin/cat' (TOK_ALIASLIST k v tkey ten one tvalue)))) (TOK_DISTRIBUTEBY (TOK_FUNCTION rand 3)) (TOK_SORTBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL tvalue)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL tkey)))))
+ (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_TAB dest1)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (TOK_ALLCOLREF src) (. (TOK_TABLE_OR_COL src) key) (TOK_FUNCTION TOK_INT (/ (. (TOK_TABLE_OR_COL src) key) 10)) (TOK_FUNCTION TOK_INT (% (. (TOK_TABLE_OR_COL src) key) 10)) (. (TOK_TABLE_OR_COL src) value)) TOK_SERDE '/bin/cat' TOK_SERDE TOK_RECORDREADER (TOK_ALIASLIST k v tkey ten one tvalue)))) (TOK_DISTRIBUTEBY (TOK_FUNCTION rand 3)) (TOK_SORTBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL tvalue)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL tkey)))))
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -111,7 +111,7 @@
Output: default/dest1
query: SELECT dest1.* FROM dest1
Input: default/dest1
-Output: file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_2/build/ql/tmp/1294545033/10000
+Output: file:/data/users/njain/hive4/hive4/build/ql/tmp/170788469/10000
0 val_0 0 0 0 val_0
0 val_0 0 0 0 val_0
0 val_0 0 0 0 val_0
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/nullscript.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/nullscript.q.out?rev=807330&r1=807329&r2=807330&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/nullscript.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/nullscript.q.out Mon Aug 24 18:23:19 2009
@@ -5,7 +5,7 @@
query: explain
select transform(key) using '/bin/cat' as key1 from nullscript
ABSTRACT SYNTAX TREE:
- (TOK_QUERY (TOK_FROM (TOK_TABREF nullscript)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (TOK_TABLE_OR_COL key)) '/bin/cat' (TOK_ALIASLIST key1))))))
+ (TOK_QUERY (TOK_FROM (TOK_TABREF nullscript)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (TOK_TABLE_OR_COL key)) TOK_SERDE '/bin/cat' TOK_SERDE TOK_RECORDREADER (TOK_ALIASLIST key1))))))
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -42,7 +42,7 @@
query: select transform(key) using '/bin/cat' as key1 from nullscript
Input: default/nullscript
-Output: file:/data/users/athusoo/commits/hive_trunk_ws1/build/ql/tmp/614707845/10000
+Output: file:/data/users/njain/hive4/hive4/build/ql/tmp/978374417/10000
238
86
311
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_transform.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_transform.q.out?rev=807330&r1=807329&r2=807330&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_transform.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_transform.q.out Mon Aug 24 18:23:19 2009
@@ -7,7 +7,7 @@
) tmap
SELECT tmap.tkey, tmap.tvalue WHERE tmap.tkey < 100
ABSTRACT SYNTAX TREE:
- (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL src) key) (. (TOK_TABLE_OR_COL src) value)) '/bin/cat' (TOK_ALIASLIST tkey tvalue)))) (TOK_CLUSTERBY (TOK_TABLE_OR_COL tkey)))) tmap)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) tkey)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) tvalue))) (TOK_WHERE (< (. (TOK_TABLE_OR_COL tmap) tkey) 100))))
+ (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL src) key) (. (TOK_TABLE_OR_COL src) value)) TOK_SERDE '/bin/cat' TOK_SERDE TOK_RECORDREADER (TOK_ALIASLIST tkey tvalue)))) (TOK_CLUSTERBY (TOK_TABLE_OR_COL tkey)))) tmap)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) tkey)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) tvalue))) (TOK_WHERE (< (. (TOK_TABLE_OR_COL tmap) tkey) 100))))
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -79,7 +79,7 @@
) tmap
SELECT tmap.tkey, tmap.tvalue WHERE tmap.tkey < 100
Input: default/src
-Output: file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_2/build/ql/tmp/603794391/10000
+Output: file:/data/users/njain/hive4/hive4/build/ql/tmp/1178089894/10000
0 val_0
0 val_0
0 val_0
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/regexp_extract.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/regexp_extract.q.out?rev=807330&r1=807329&r2=807330&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/regexp_extract.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/regexp_extract.q.out Mon Aug 24 18:23:19 2009
@@ -7,7 +7,7 @@
) tmap
SELECT tmap.key, regexp_extract(tmap.value, 'val_(\\d+\\t\\d+)',1) WHERE tmap.key < 100
ABSTRACT SYNTAX TREE:
- (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL src) key) (. (TOK_TABLE_OR_COL src) value) (+ 1 2) (+ 3 4)) '/bin/cat'))) (TOK_CLUSTERBY (TOK_TABLE_OR_COL key)))) tmap)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) key)) (TOK_SELEXPR (TOK_FUNCTION regexp_extract (. (TOK_TABLE_OR_COL tmap) value) 'val_(\\d+\\t\\d+)' 1))) (TOK_WHERE (< (. (TOK_TABLE_OR_COL tmap) key) 100))))
+ (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL src) key) (. (TOK_TABLE_OR_COL src) value) (+ 1 2) (+ 3 4)) TOK_SERDE '/bin/cat' TOK_SERDE TOK_RECORDREADER))) (TOK_CLUSTERBY (TOK_TABLE_OR_COL key)))) tmap)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) key)) (TOK_SELEXPR (TOK_FUNCTION regexp_extract (. (TOK_TABLE_OR_COL tmap) value) 'val_(\\d+\\t\\d+)' 1))) (TOK_WHERE (< (. (TOK_TABLE_OR_COL tmap) key) 100))))
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -39,6 +39,7 @@
properties:
columns key,value
serialization.format 9
+ columns.types string,string
serialization.last.column.takes.rest true
Reduce Output Operator
key expressions:
@@ -56,9 +57,9 @@
type: string
Needs Tagging: false
Path -> Alias:
- file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_2/build/ql/test/data/warehouse/src [tmap:src]
+ file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/src [tmap:src]
Path -> Partition:
- file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_2/build/ql/test/data/warehouse/src
+ file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/src
Partition
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -73,7 +74,7 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_2/build/ql/test/data/warehouse/src
+ location file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/src
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: src
Reduce Operator Tree:
@@ -93,7 +94,7 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_2/build/ql/tmp/1993505599/10001
+ directory: file:/data/users/njain/hive4/hive4/build/ql/tmp/760693203/10001
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -106,6 +107,7 @@
Fetch Operator
limit: -1
+
query: FROM (
FROM src
SELECT TRANSFORM(src.key, src.value, 1+2, 3+4)
@@ -114,7 +116,7 @@
) tmap
SELECT tmap.key, regexp_extract(tmap.value, 'val_(\\d+\\t\\d+)',1) WHERE tmap.key < 100
Input: default/src
-Output: file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_2/build/ql/tmp/211175819/10000
+Output: file:/data/users/njain/hive4/hive4/build/ql/tmp/940676220/10000
0 0 3
0 0 3
0 0 3
@@ -208,7 +210,7 @@
) tmap
SELECT tmap.key, regexp_extract(tmap.value, 'val_(\\d+\\t\\d+)') WHERE tmap.key < 100
ABSTRACT SYNTAX TREE:
- (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL src) key) (. (TOK_TABLE_OR_COL src) value) (+ 1 2) (+ 3 4)) '/bin/cat'))) (TOK_CLUSTERBY (TOK_TABLE_OR_COL key)))) tmap)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) key)) (TOK_SELEXPR (TOK_FUNCTION regexp_extract (. (TOK_TABLE_OR_COL tmap) value) 'val_(\\d+\\t\\d+)'))) (TOK_WHERE (< (. (TOK_TABLE_OR_COL tmap) key) 100))))
+ (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL src) key) (. (TOK_TABLE_OR_COL src) value) (+ 1 2) (+ 3 4)) TOK_SERDE '/bin/cat' TOK_SERDE TOK_RECORDREADER))) (TOK_CLUSTERBY (TOK_TABLE_OR_COL key)))) tmap)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) key)) (TOK_SELEXPR (TOK_FUNCTION regexp_extract (. (TOK_TABLE_OR_COL tmap) value) 'val_(\\d+\\t\\d+)'))) (TOK_WHERE (< (. (TOK_TABLE_OR_COL tmap) key) 100))))
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -240,6 +242,7 @@
properties:
columns key,value
serialization.format 9
+ columns.types string,string
serialization.last.column.takes.rest true
Reduce Output Operator
key expressions:
@@ -257,9 +260,9 @@
type: string
Needs Tagging: false
Path -> Alias:
- file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_2/build/ql/test/data/warehouse/src [tmap:src]
+ file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/src [tmap:src]
Path -> Partition:
- file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_2/build/ql/test/data/warehouse/src
+ file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/src
Partition
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -274,7 +277,7 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_2/build/ql/test/data/warehouse/src
+ location file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/src
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: src
Reduce Operator Tree:
@@ -294,7 +297,7 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_2/build/ql/tmp/97566241/10001
+ directory: file:/data/users/njain/hive4/hive4/build/ql/tmp/1376686239/10001
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -307,6 +310,7 @@
Fetch Operator
limit: -1
+
query: FROM (
FROM src
SELECT TRANSFORM(src.key, src.value, 1+2, 3+4)
@@ -315,7 +319,7 @@
) tmap
SELECT tmap.key, regexp_extract(tmap.value, 'val_(\\d+\\t\\d+)') WHERE tmap.key < 100
Input: default/src
-Output: file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_2/build/ql/tmp/1666620483/10000
+Output: file:/data/users/njain/hive4/hive4/build/ql/tmp/2047565787/10000
0 0 3
0 0 3
0 0 3
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/transform_ppr1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/transform_ppr1.q.out?rev=807330&r1=807329&r2=807330&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/transform_ppr1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/transform_ppr1.q.out Mon Aug 24 18:23:19 2009
@@ -7,7 +7,7 @@
) tmap
SELECT tmap.tkey, tmap.tvalue WHERE tmap.tkey < 100 AND tmap.ds = '2008-04-08'
ABSTRACT SYNTAX TREE:
- (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF srcpart src)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL src) ds) (. (TOK_TABLE_OR_COL src) key) (. (TOK_TABLE_OR_COL src) value)) '/bin/cat' (TOK_ALIASLIST ds tkey tvalue)))) (TOK_CLUSTERBY (TOK_TABLE_OR_COL tkey)))) tmap)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) tkey)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) tvalue))) (TOK_WHERE (AND (< (. (TOK_TABLE_OR_COL tmap) tkey) 100) (= (. (TOK_TABLE_OR_COL tmap) ds) '2008-04-08')))))
+ (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF srcpart src)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL src) ds) (. (TOK_TABLE_OR_COL src) key) (. (TOK_TABLE_OR_COL src) value)) TOK_SERDE '/bin/cat' TOK_SERDE TOK_RECORDREADER (TOK_ALIASLIST ds tkey tvalue)))) (TOK_CLUSTERBY (TOK_TABLE_OR_COL tkey)))) tmap)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) tkey)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) tvalue))) (TOK_WHERE (AND (< (. (TOK_TABLE_OR_COL tmap) tkey) 100) (= (. (TOK_TABLE_OR_COL tmap) ds) '2008-04-08')))))
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -37,6 +37,7 @@
properties:
columns ds,tkey,tvalue
serialization.format 9
+ columns.types string,string,string
Reduce Output Operator
key expressions:
expr: tkey
@@ -55,12 +56,12 @@
type: string
Needs Tagging: false
Path -> Alias:
- file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [tmap:src]
- file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [tmap:src]
- file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11 [tmap:src]
- file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12 [tmap:src]
+ file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [tmap:src]
+ file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [tmap:src]
+ file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11 [tmap:src]
+ file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12 [tmap:src]
Path -> Partition:
- file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
+ file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
Partition
partition values:
ds 2008-04-08
@@ -79,10 +80,10 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_0/build/ql/test/data/warehouse/srcpart
+ location file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcpart
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
- file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
+ file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
Partition
partition values:
ds 2008-04-08
@@ -101,10 +102,10 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_0/build/ql/test/data/warehouse/srcpart
+ location file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcpart
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
- file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11
+ file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11
Partition
partition values:
ds 2008-04-09
@@ -123,10 +124,10 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_0/build/ql/test/data/warehouse/srcpart
+ location file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcpart
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
- file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12
+ file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12
Partition
partition values:
ds 2008-04-09
@@ -145,7 +146,7 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_0/build/ql/test/data/warehouse/srcpart
+ location file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcpart
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
Reduce Operator Tree:
@@ -165,7 +166,7 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_0/build/ql/tmp/1007016936/10001
+ directory: file:/data/users/njain/hive4/hive4/build/ql/tmp/1048025632/10001
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -190,7 +191,7 @@
Input: default/srcpart/ds=2008-04-08/hr=12
Input: default/srcpart/ds=2008-04-09/hr=11
Input: default/srcpart/ds=2008-04-09/hr=12
-Output: file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_0/build/ql/tmp/312576026/10000
+Output: file:/data/users/njain/hive4/hive4/build/ql/tmp/1138294576/10000
0 val_0
0 val_0
0 val_0
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/transform_ppr2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/transform_ppr2.q.out?rev=807330&r1=807329&r2=807330&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/transform_ppr2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/transform_ppr2.q.out Mon Aug 24 18:23:19 2009
@@ -8,7 +8,7 @@
) tmap
SELECT tmap.tkey, tmap.tvalue WHERE tmap.tkey < 100
ABSTRACT SYNTAX TREE:
- (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF srcpart src)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL src) ds) (. (TOK_TABLE_OR_COL src) key) (. (TOK_TABLE_OR_COL src) value)) '/bin/cat' (TOK_ALIASLIST ds tkey tvalue)))) (TOK_WHERE (= (. (TOK_TABLE_OR_COL src) ds) '2008-04-08')) (TOK_CLUSTERBY (TOK_TABLE_OR_COL tkey)))) tmap)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) tkey)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) tvalue))) (TOK_WHERE (< (. (TOK_TABLE_OR_COL tmap) tkey) 100))))
+ (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF srcpart src)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL src) ds) (. (TOK_TABLE_OR_COL src) key) (. (TOK_TABLE_OR_COL src) value)) TOK_SERDE '/bin/cat' TOK_SERDE TOK_RECORDREADER (TOK_ALIASLIST ds tkey tvalue)))) (TOK_WHERE (= (. (TOK_TABLE_OR_COL src) ds) '2008-04-08')) (TOK_CLUSTERBY (TOK_TABLE_OR_COL tkey)))) tmap)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) tkey)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) tvalue))) (TOK_WHERE (< (. (TOK_TABLE_OR_COL tmap) tkey) 100))))
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -48,6 +48,7 @@
properties:
columns ds,tkey,tvalue
serialization.format 9
+ columns.types string,string,string
Reduce Output Operator
key expressions:
expr: tkey
@@ -66,10 +67,10 @@
type: string
Needs Tagging: false
Path -> Alias:
- file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_1/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [tmap:src]
- file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_1/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [tmap:src]
+ file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [tmap:src]
+ file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [tmap:src]
Path -> Partition:
- file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_1/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
+ file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
Partition
partition values:
ds 2008-04-08
@@ -88,10 +89,10 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_1/build/ql/test/data/warehouse/srcpart
+ location file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcpart
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
- file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_1/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
+ file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
Partition
partition values:
ds 2008-04-08
@@ -110,7 +111,7 @@
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_1/build/ql/test/data/warehouse/srcpart
+ location file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/srcpart
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
Reduce Operator Tree:
@@ -130,7 +131,7 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_1/build/ql/tmp/938544971/10001
+ directory: file:/data/users/njain/hive4/hive4/build/ql/tmp/495053440/10001
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -154,7 +155,7 @@
SELECT tmap.tkey, tmap.tvalue WHERE tmap.tkey < 100
Input: default/srcpart/ds=2008-04-08/hr=11
Input: default/srcpart/ds=2008-04-08/hr=12
-Output: file:/data/users/zshao/tools/699-trunk-apache-hive/.ptest_1/build/ql/tmp/1145900420/10000
+Output: file:/data/users/njain/hive4/hive4/build/ql/tmp/211417168/10000
0 val_0
0 val_0
0 val_0
Modified: hadoop/hive/trunk/ql/src/test/results/compiler/parse/input20.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/compiler/parse/input20.q.out?rev=807330&r1=807329&r2=807330&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/compiler/parse/input20.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/compiler/parse/input20.q.out Mon Aug 24 18:23:19 2009
@@ -1 +1 @@
-(TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (% (. (TOK_TABLE_OR_COL src) key) 2) (% (. (TOK_TABLE_OR_COL src) key) 5)) 'cat'))) (TOK_CLUSTERBY (TOK_TABLE_OR_COL key)))) tmap)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL tmap) key) (. (TOK_TABLE_OR_COL tmap) value)) 'uniq -c | sed "s@^ *@@" | sed "s@\t@_@" | sed "s@ @\t@"' (TOK_ALIASLIST key value)))))) null
\ No newline at end of file
+(TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (% (. (TOK_TABLE_OR_COL src) key) 2) (% (. (TOK_TABLE_OR_COL src) key) 5)) TOK_SERDE 'cat' TOK_SERDE TOK_RECORDREADER))) (TOK_CLUSTERBY (TOK_TABLE_OR_COL key)))) tmap)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL tmap) key) (. (TOK_TABLE_OR_COL tmap) value)) TOK_SERDE 'uniq -c | sed "s@^ *@@" | sed "s@\t@_@" | sed "s@ @\t@"' TOK_SERDE TOK_RECORDREADER (TOK_ALIASLIST key value)))))) null
\ No newline at end of file
Modified: hadoop/hive/trunk/ql/src/test/results/compiler/parse/input4.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/compiler/parse/input4.q.out?rev=807330&r1=807329&r2=807330&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/compiler/parse/input4.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/compiler/parse/input4.q.out Mon Aug 24 18:23:19 2009
@@ -1 +1 @@
-(TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL src) key) (. (TOK_TABLE_OR_COL src) value)) '/bin/cat' (TOK_ALIASLIST tkey tvalue)))) (TOK_CLUSTERBY (TOK_TABLE_OR_COL tkey)))) tmap)) (TOK_INSERT (TOK_DESTINATION (TOK_TAB dest1)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) tkey)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) tvalue))) (TOK_WHERE (< (. (TOK_TABLE_OR_COL tmap) tkey) 100)))) null
\ No newline at end of file
+(TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL src) key) (. (TOK_TABLE_OR_COL src) value)) TOK_SERDE '/bin/cat' TOK_SERDE TOK_RECORDREADER (TOK_ALIASLIST tkey tvalue)))) (TOK_CLUSTERBY (TOK_TABLE_OR_COL tkey)))) tmap)) (TOK_INSERT (TOK_DESTINATION (TOK_TAB dest1)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) tkey)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) tvalue))) (TOK_WHERE (< (. (TOK_TABLE_OR_COL tmap) tkey) 100)))) null
\ No newline at end of file
Modified: hadoop/hive/trunk/ql/src/test/results/compiler/parse/input5.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/compiler/parse/input5.q.out?rev=807330&r1=807329&r2=807330&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/compiler/parse/input5.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/compiler/parse/input5.q.out Mon Aug 24 18:23:19 2009
@@ -1 +1 @@
-(TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF src_thrift)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL src_thrift) lint) (. (TOK_TABLE_OR_COL src_thrift) lintstring)) '/bin/cat' (TOK_ALIASLIST tkey tvalue)))) (TOK_CLUSTERBY (TOK_TABLE_OR_COL tkey)))) tmap)) (TOK_INSERT (TOK_DESTINATION (TOK_TAB dest1)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) tkey)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) tvalue))))) null
\ No newline at end of file
+(TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF src_thrift)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL src_thrift) lint) (. (TOK_TABLE_OR_COL src_thrift) lintstring)) TOK_SERDE '/bin/cat' TOK_SERDE TOK_RECORDREADER (TOK_ALIASLIST tkey tvalue)))) (TOK_CLUSTERBY (TOK_TABLE_OR_COL tkey)))) tmap)) (TOK_INSERT (TOK_DESTINATION (TOK_TAB dest1)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) tkey)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL tmap) tvalue))))) null
\ No newline at end of file
Modified: hadoop/hive/trunk/ql/src/test/results/compiler/plan/input20.q.xml
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/compiler/plan/input20.q.xml?rev=807330&r1=807329&r2=807330&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/compiler/plan/input20.q.xml (original)
+++ hadoop/hive/trunk/ql/src/test/results/compiler/plan/input20.q.xml Mon Aug 24 18:23:19 2009
@@ -64,7 +64,7 @@
</void>
<void method="put">
<string>location</string>
- <string>file:/data/users/njain/hive3/hive3/build/ql/test/data/warehouse/src</string>
+ <string>file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/src</string>
</void>
</object>
</void>
@@ -303,6 +303,9 @@
</void>
<void property="conf">
<object class="org.apache.hadoop.hive.ql.plan.scriptDesc">
+ <void property="outRecordReaderClass">
+ <class>org.apache.hadoop.hive.ql.exec.TextRecordReader</class>
+ </void>
<void property="scriptCmd">
<string>cat</string>
</void>
@@ -321,12 +324,20 @@
<object class="java.util.Properties">
<void method="put">
<string>columns</string>
- <string></string>
+ <string>_col0,_col1</string>
+ </void>
+ <void method="put">
+ <string>serialization.use.json.object</string>
+ <string>true</string>
</void>
<void method="put">
<string>serialization.format</string>
<string>9</string>
</void>
+ <void method="put">
+ <string>columns.types</string>
+ <string>double,double</string>
+ </void>
</object>
</void>
</object>
@@ -353,6 +364,10 @@
<string>9</string>
</void>
<void method="put">
+ <string>columns.types</string>
+ <string>string,string</string>
+ </void>
+ <void method="put">
<string>serialization.last.column.takes.rest</string>
<string>true</string>
</void>
@@ -654,7 +669,7 @@
<void property="pathToAliases">
<object class="java.util.LinkedHashMap">
<void method="put">
- <string>file:/data/users/njain/hive3/hive3/build/ql/test/data/warehouse/src</string>
+ <string>file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/src</string>
<object class="java.util.ArrayList">
<void method="add">
<string>tmap:src</string>
@@ -666,7 +681,7 @@
<void property="pathToPartitionInfo">
<object class="java.util.LinkedHashMap">
<void method="put">
- <string>file:/data/users/njain/hive3/hive3/build/ql/test/data/warehouse/src</string>
+ <string>file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/src</string>
<object class="org.apache.hadoop.hive.ql.plan.partitionDesc">
<void property="partSpec">
<object idref="LinkedHashMap0"/>
@@ -711,7 +726,7 @@
<void property="conf">
<object class="org.apache.hadoop.hive.ql.plan.fileSinkDesc">
<void property="dirName">
- <string>file:/data/users/njain/hive3/hive3/build/ql/tmp/903491729/10001</string>
+ <string>file:/data/users/njain/hive4/hive4/build/ql/tmp/1544350332/10001</string>
</void>
<void property="tableInfo">
<object class="org.apache.hadoop.hive.ql.plan.tableDesc">
@@ -801,6 +816,9 @@
</void>
<void property="conf">
<object class="org.apache.hadoop.hive.ql.plan.scriptDesc">
+ <void property="outRecordReaderClass">
+ <class>org.apache.hadoop.hive.ql.exec.TextRecordReader</class>
+ </void>
<void property="scriptCmd">
<string>uniq -c | sed "s@^ *@@" | sed "s@\t@_@" | sed "s@ @\t@"</string>
</void>
@@ -819,12 +837,20 @@
<object class="java.util.Properties">
<void method="put">
<string>columns</string>
- <string></string>
+ <string>_col0,_col1</string>
+ </void>
+ <void method="put">
+ <string>serialization.use.json.object</string>
+ <string>true</string>
</void>
<void method="put">
<string>serialization.format</string>
<string>9</string>
</void>
+ <void method="put">
+ <string>columns.types</string>
+ <string>string,string</string>
+ </void>
</object>
</void>
</object>
@@ -850,6 +876,10 @@
<string>serialization.format</string>
<string>9</string>
</void>
+ <void method="put">
+ <string>columns.types</string>
+ <string>string,string</string>
+ </void>
</object>
</void>
</object>
Modified: hadoop/hive/trunk/ql/src/test/results/compiler/plan/input4.q.xml
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/compiler/plan/input4.q.xml?rev=807330&r1=807329&r2=807330&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/compiler/plan/input4.q.xml (original)
+++ hadoop/hive/trunk/ql/src/test/results/compiler/plan/input4.q.xml Mon Aug 24 18:23:19 2009
@@ -26,7 +26,7 @@
<boolean>true</boolean>
</void>
<void property="sourceDir">
- <string>file:/data/users/njain/hive3/hive3/build/ql/tmp/1591943582/10000</string>
+ <string>file:/data/users/njain/hive4/hive4/build/ql/tmp/1849422817/10000</string>
</void>
<void property="table">
<object id="tableDesc0" class="org.apache.hadoop.hive.ql.plan.tableDesc">
@@ -79,7 +79,7 @@
</void>
<void method="put">
<string>location</string>
- <string>file:/data/users/njain/hive3/hive3/build/ql/test/data/warehouse/dest1</string>
+ <string>file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/dest1</string>
</void>
</object>
</void>
@@ -89,7 +89,7 @@
</object>
</void>
<void property="tmpDir">
- <string>file:/data/users/njain/hive3/hive3/build/ql/tmp/1591943582/10001</string>
+ <string>file:/data/users/njain/hive4/hive4/build/ql/tmp/1849422817/10001</string>
</void>
</object>
</void>
@@ -162,7 +162,7 @@
</void>
<void method="put">
<string>location</string>
- <string>file:/data/users/njain/hive3/hive3/build/ql/test/data/warehouse/src</string>
+ <string>file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/src</string>
</void>
</object>
</void>
@@ -401,6 +401,9 @@
</void>
<void property="conf">
<object class="org.apache.hadoop.hive.ql.plan.scriptDesc">
+ <void property="outRecordReaderClass">
+ <class>org.apache.hadoop.hive.ql.exec.TextRecordReader</class>
+ </void>
<void property="scriptCmd">
<string>/bin/cat</string>
</void>
@@ -419,12 +422,20 @@
<object class="java.util.Properties">
<void method="put">
<string>columns</string>
- <string></string>
+ <string>_col0,_col1</string>
+ </void>
+ <void method="put">
+ <string>serialization.use.json.object</string>
+ <string>true</string>
</void>
<void method="put">
<string>serialization.format</string>
<string>9</string>
</void>
+ <void method="put">
+ <string>columns.types</string>
+ <string>string,string</string>
+ </void>
</object>
</void>
</object>
@@ -450,6 +461,10 @@
<string>serialization.format</string>
<string>9</string>
</void>
+ <void method="put">
+ <string>columns.types</string>
+ <string>string,string</string>
+ </void>
</object>
</void>
</object>
@@ -675,7 +690,7 @@
<void property="pathToAliases">
<object class="java.util.LinkedHashMap">
<void method="put">
- <string>file:/data/users/njain/hive3/hive3/build/ql/test/data/warehouse/src</string>
+ <string>file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/src</string>
<object class="java.util.ArrayList">
<void method="add">
<string>tmap:src</string>
@@ -687,7 +702,7 @@
<void property="pathToPartitionInfo">
<object class="java.util.LinkedHashMap">
<void method="put">
- <string>file:/data/users/njain/hive3/hive3/build/ql/test/data/warehouse/src</string>
+ <string>file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/src</string>
<object class="org.apache.hadoop.hive.ql.plan.partitionDesc">
<void property="partSpec">
<object idref="LinkedHashMap0"/>
@@ -735,7 +750,7 @@
<int>1</int>
</void>
<void property="dirName">
- <string>file:/data/users/njain/hive3/hive3/build/ql/tmp/1591943582/10000</string>
+ <string>file:/data/users/njain/hive4/hive4/build/ql/tmp/1849422817/10000</string>
</void>
<void property="tableInfo">
<object idref="tableDesc0"/>
Modified: hadoop/hive/trunk/ql/src/test/results/compiler/plan/input5.q.xml
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/compiler/plan/input5.q.xml?rev=807330&r1=807329&r2=807330&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/compiler/plan/input5.q.xml (original)
+++ hadoop/hive/trunk/ql/src/test/results/compiler/plan/input5.q.xml Mon Aug 24 18:23:19 2009
@@ -26,7 +26,7 @@
<boolean>true</boolean>
</void>
<void property="sourceDir">
- <string>file:/data/users/njain/hive3/hive3/build/ql/tmp/375153903/10000</string>
+ <string>file:/data/users/njain/hive4/hive4/build/ql/tmp/300198296/10000</string>
</void>
<void property="table">
<object id="tableDesc0" class="org.apache.hadoop.hive.ql.plan.tableDesc">
@@ -79,7 +79,7 @@
</void>
<void method="put">
<string>location</string>
- <string>file:/data/users/njain/hive3/hive3/build/ql/test/data/warehouse/dest1</string>
+ <string>file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/dest1</string>
</void>
</object>
</void>
@@ -89,7 +89,7 @@
</object>
</void>
<void property="tmpDir">
- <string>file:/data/users/njain/hive3/hive3/build/ql/tmp/375153903/10001</string>
+ <string>file:/data/users/njain/hive4/hive4/build/ql/tmp/300198296/10001</string>
</void>
</object>
</void>
@@ -166,7 +166,7 @@
</void>
<void method="put">
<string>location</string>
- <string>file:/data/users/njain/hive3/hive3/build/ql/test/data/warehouse/src_thrift</string>
+ <string>file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/src_thrift</string>
</void>
</object>
</void>
@@ -405,6 +405,9 @@
</void>
<void property="conf">
<object class="org.apache.hadoop.hive.ql.plan.scriptDesc">
+ <void property="outRecordReaderClass">
+ <class>org.apache.hadoop.hive.ql.exec.TextRecordReader</class>
+ </void>
<void property="scriptCmd">
<string>/bin/cat</string>
</void>
@@ -423,12 +426,20 @@
<object class="java.util.Properties">
<void method="put">
<string>columns</string>
- <string></string>
+ <string>_col0,_col1</string>
+ </void>
+ <void method="put">
+ <string>serialization.use.json.object</string>
+ <string>true</string>
</void>
<void method="put">
<string>serialization.format</string>
<string>9</string>
</void>
+ <void method="put">
+ <string>columns.types</string>
+ <string>array<int>,array<struct<myint:int,mystring:string,underscore_int:int>></string>
+ </void>
</object>
</void>
</object>
@@ -454,6 +465,10 @@
<string>serialization.format</string>
<string>9</string>
</void>
+ <void method="put">
+ <string>columns.types</string>
+ <string>string,string</string>
+ </void>
</object>
</void>
</object>
@@ -742,7 +757,7 @@
<void property="pathToAliases">
<object class="java.util.LinkedHashMap">
<void method="put">
- <string>file:/data/users/njain/hive3/hive3/build/ql/test/data/warehouse/src_thrift</string>
+ <string>file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/src_thrift</string>
<object class="java.util.ArrayList">
<void method="add">
<string>tmap:src_thrift</string>
@@ -754,7 +769,7 @@
<void property="pathToPartitionInfo">
<object class="java.util.LinkedHashMap">
<void method="put">
- <string>file:/data/users/njain/hive3/hive3/build/ql/test/data/warehouse/src_thrift</string>
+ <string>file:/data/users/njain/hive4/hive4/build/ql/test/data/warehouse/src_thrift</string>
<object class="org.apache.hadoop.hive.ql.plan.partitionDesc">
<void property="partSpec">
<object idref="LinkedHashMap0"/>
@@ -798,7 +813,7 @@
<int>1</int>
</void>
<void property="dirName">
- <string>file:/data/users/njain/hive3/hive3/build/ql/tmp/375153903/10000</string>
+ <string>file:/data/users/njain/hive4/hive4/build/ql/tmp/300198296/10000</string>
</void>
<void property="tableInfo">
<object idref="tableDesc0"/>
Modified: hadoop/hive/trunk/serde/if/serde.thrift
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/serde/if/serde.thrift?rev=807330&r1=807329&r2=807330&view=diff
==============================================================================
--- hadoop/hive/trunk/serde/if/serde.thrift (original)
+++ hadoop/hive/trunk/serde/if/serde.thrift Mon Aug 24 18:23:19 2009
@@ -12,6 +12,7 @@
const string SERIALIZATION_NULL_FORMAT = "serialization.null.format"
const string SERIALIZATION_LAST_COLUMN_TAKES_REST = "serialization.last.column.takes.rest"
const string SERIALIZATION_SORT_ORDER = "serialization.sort.order"
+const string SERIALIZATION_USE_JSON_OBJECTS = "serialization.use.json.object"
const string FIELD_DELIM = "field.delim"
const string COLLECTION_DELIM = "colelction.delim"
Modified: hadoop/hive/trunk/serde/src/gen-java/org/apache/hadoop/hive/serde/Constants.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/serde/src/gen-java/org/apache/hadoop/hive/serde/Constants.java?rev=807330&r1=807329&r2=807330&view=diff
==============================================================================
--- hadoop/hive/trunk/serde/src/gen-java/org/apache/hadoop/hive/serde/Constants.java (original)
+++ hadoop/hive/trunk/serde/src/gen-java/org/apache/hadoop/hive/serde/Constants.java Mon Aug 24 18:23:19 2009
@@ -30,6 +30,8 @@
public static final String SERIALIZATION_SORT_ORDER = "serialization.sort.order";
+ public static final String SERIALIZATION_USE_JSON_OBJECTS = "serialization.use.json.object";
+
public static final String FIELD_DELIM = "field.delim";
public static final String COLLECTION_DELIM = "colelction.delim";
Modified: hadoop/hive/trunk/serde/src/gen-php/serde_constants.php
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/serde/src/gen-php/serde_constants.php?rev=807330&r1=807329&r2=807330&view=diff
==============================================================================
--- hadoop/hive/trunk/serde/src/gen-php/serde_constants.php (original)
+++ hadoop/hive/trunk/serde/src/gen-php/serde_constants.php Mon Aug 24 18:23:19 2009
@@ -22,6 +22,8 @@
$GLOBALS['serde_CONSTANTS']['SERIALIZATION_SORT_ORDER'] = "serialization.sort.order";
+$GLOBALS['serde_CONSTANTS']['SERIALIZATION_USE_JSON_OBJECTS'] = "serialization.use.json.object";
+
$GLOBALS['serde_CONSTANTS']['FIELD_DELIM'] = "field.delim";
$GLOBALS['serde_CONSTANTS']['COLLECTION_DELIM'] = "colelction.delim";
Modified: hadoop/hive/trunk/serde/src/gen-py/org_apache_hadoop_hive_serde/constants.py
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/serde/src/gen-py/org_apache_hadoop_hive_serde/constants.py?rev=807330&r1=807329&r2=807330&view=diff
==============================================================================
--- hadoop/hive/trunk/serde/src/gen-py/org_apache_hadoop_hive_serde/constants.py (original)
+++ hadoop/hive/trunk/serde/src/gen-py/org_apache_hadoop_hive_serde/constants.py Mon Aug 24 18:23:19 2009
@@ -21,6 +21,8 @@
SERIALIZATION_SORT_ORDER = "serialization.sort.order"
+SERIALIZATION_USE_JSON_OBJECTS = "serialization.use.json.object"
+
FIELD_DELIM = "field.delim"
COLLECTION_DELIM = "colelction.delim"
Modified: hadoop/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/binarysortable/InputByteBuffer.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/binarysortable/InputByteBuffer.java?rev=807330&r1=807329&r2=807330&view=diff
==============================================================================
--- hadoop/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/binarysortable/InputByteBuffer.java (original)
+++ hadoop/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/binarysortable/InputByteBuffer.java Mon Aug 24 18:23:19 2009
@@ -39,6 +39,10 @@
this.end = end;
}
+ public final byte read() throws IOException {
+ return read(false);
+ }
+
/**
* Read one byte from the byte buffer.
* Final method to help inlining.
@@ -71,6 +75,10 @@
start = position;
}
+ public final int getEnd() {
+ return end;
+ }
+
/**
* Returns the underlying byte array.
*/
Modified: hadoop/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/binarysortable/OutputByteBuffer.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/binarysortable/OutputByteBuffer.java?rev=807330&r1=807329&r2=807330&view=diff
==============================================================================
--- hadoop/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/binarysortable/OutputByteBuffer.java (original)
+++ hadoop/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/binarysortable/OutputByteBuffer.java Mon Aug 24 18:23:19 2009
@@ -35,6 +35,10 @@
length = 0;
}
+ public final void write(byte b) {
+ write(b, false);
+ }
+
/**
* Write one byte to the byte buffer.
* Final method to help inlining.
Modified: hadoop/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazySimpleSerDe.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazySimpleSerDe.java?rev=807330&r1=807329&r2=807330&view=diff
==============================================================================
--- hadoop/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazySimpleSerDe.java (original)
+++ hadoop/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazySimpleSerDe.java Mon Aug 24 18:23:19 2009
@@ -112,6 +112,7 @@
boolean escaped;
byte escapeChar;
boolean[] needsEscape;
+ boolean jsonSerialize;
public List<TypeInfo> getColumnTypes() {
return columnTypes;
@@ -150,6 +151,21 @@
public boolean[] getNeedsEscape() {
return needsEscape;
}
+
+ /**
+ * @return the jsonSerialize
+ */
+ public boolean isJsonSerialize() {
+ return jsonSerialize;
+ }
+
+ /**
+ * @param jsonSerialize the jsonSerialize to set
+ */
+ public void setJsonSerialize(boolean jsonSerialize) {
+ this.jsonSerialize = jsonSerialize;
+ }
+
}
SerDeParameters serdeParams = null;
@@ -176,6 +192,9 @@
serdeParams.isEscaped(),
serdeParams.getEscapeChar());
+ if (serdeParams.isJsonSerialize())
+ setUseJSONSerialize(true);
+
cachedLazyStruct = (LazyStruct) LazyFactory
.createLazyObject(cachedObjectInspector);
@@ -212,6 +231,10 @@
serdeParams.lastColumnTakesRest = (lastColumnTakesRestString != null && lastColumnTakesRestString
.equalsIgnoreCase("true"));
+ String useJsonSerialize = tbl
+ .getProperty(Constants.SERIALIZATION_USE_JSON_OBJECTS);
+ serdeParams.jsonSerialize = (useJsonSerialize != null && useJsonSerialize.equalsIgnoreCase("true"));
+
// Read the configuration parameters
String columnNameProperty = tbl.getProperty(Constants.LIST_COLUMNS);
// NOTE: if "columns.types" is missing, all columns will be of String type