You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by am...@apache.org on 2013/04/05 12:34:11 UTC
svn commit: r1464915 [4/5] - in /hive/branches/HIVE-4115: ./
common/src/java/org/apache/hadoop/hive/common/
common/src/java/org/apache/hadoop/hive/conf/ conf/ data/files/
hcatalog/build-support/ant/ hcatalog/historical/branches/
hcatalog/historical/sit...
Modified: hive/branches/HIVE-4115/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcFile.java
URL: http://svn.apache.org/viewvc/hive/branches/HIVE-4115/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcFile.java?rev=1464915&r1=1464914&r2=1464915&view=diff
==============================================================================
--- hive/branches/HIVE-4115/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcFile.java (original)
+++ hive/branches/HIVE-4115/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcFile.java Fri Apr 5 10:34:08 2013
@@ -656,7 +656,7 @@ public class TestOrcFile {
assertEquals(true, rows.hasNext());
row = (OrcStruct) rows.next(null);
inspector = reader.getObjectInspector();
- assertEquals("struct<time:timestamp,union:union{int, string}>",
+ assertEquals("struct<time:timestamp,union:uniontype<int,string>>",
inspector.getTypeName());
assertEquals(Timestamp.valueOf("2000-03-12 15:00:00"),
row.getFieldValue(0));
Modified: hive/branches/HIVE-4115/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcStruct.java
URL: http://svn.apache.org/viewvc/hive/branches/HIVE-4115/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcStruct.java?rev=1464915&r1=1464914&r2=1464915&view=diff
==============================================================================
--- hive/branches/HIVE-4115/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcStruct.java (original)
+++ hive/branches/HIVE-4115/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcStruct.java Fri Apr 5 10:34:08 2013
@@ -77,7 +77,7 @@ public class TestOrcStruct {
OrcStruct.createObjectInspector(typeInfo);
assertEquals("struct<c1:boolean,c2:tinyint,c3:smallint,c4:int,c5:" +
"bigint,c6:float,c7:double,c8:binary,c9:string,c10:struct<" +
- "c1:int>,c11:map<int,int>,c12:union{int},c13:array<timestamp>>",
+ "c1:int>,c11:map<int,int>,c12:uniontype<int>,c13:array<timestamp>>",
inspector.getTypeName());
assertEquals(null,
inspector.getAllStructFieldRefs().get(0).getFieldComment());
Modified: hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/columnstats_tbllvl.q
URL: http://svn.apache.org/viewvc/hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/columnstats_tbllvl.q?rev=1464915&r1=1464914&r2=1464915&view=diff
==============================================================================
--- hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/columnstats_tbllvl.q (original)
+++ hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/columnstats_tbllvl.q Fri Apr 5 10:34:08 2013
@@ -23,3 +23,16 @@ analyze table UserVisits_web_text_none c
analyze table UserVisits_web_text_none compute statistics for columns sourceIP, avgTimeOnSite, adRevenue;
+CREATE TABLE empty_tab(
+ a int,
+ b double,
+ c string,
+ d boolean,
+ e binary)
+row format delimited fields terminated by '|' stored as textfile;
+
+explain
+analyze table empty_tab compute statistics for columns a,b,c,d,e;
+
+analyze table empty_tab compute statistics for columns a,b,c,d,e;
+
Modified: hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/nonreserved_keywords_insert_into1.q
URL: http://svn.apache.org/viewvc/hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/nonreserved_keywords_insert_into1.q?rev=1464915&r1=1464914&r2=1464915&view=diff
==============================================================================
--- hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/nonreserved_keywords_insert_into1.q (original)
+++ hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/nonreserved_keywords_insert_into1.q Fri Apr 5 10:34:08 2013
@@ -1,6 +1,6 @@
DROP TABLE insert;
-CREATE TABLE insert (from INT, as STRING);
+CREATE TABLE insert (key INT, as STRING);
EXPLAIN INSERT INTO TABLE insert SELECT * FROM src LIMIT 100;
INSERT INTO TABLE insert SELECT * FROM src LIMIT 100;
Modified: hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/serde_regex.q
URL: http://svn.apache.org/viewvc/hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/serde_regex.q?rev=1464915&r1=1464914&r2=1464915&view=diff
==============================================================================
--- hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/serde_regex.q (original)
+++ hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/serde_regex.q Fri Apr 5 10:34:08 2013
@@ -39,3 +39,28 @@ SELECT * FROM serde_regex ORDER BY time;
SELECT host, size, status, time from serde_regex ORDER BY time;
DROP TABLE serde_regex;
+
+EXPLAIN
+CREATE TABLE serde_regex1(
+ key decimal,
+ value int)
+ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.RegexSerDe'
+WITH SERDEPROPERTIES (
+ "input.regex" = "([^ ]*) ([^ ]*)"
+)
+STORED AS TEXTFILE;
+
+CREATE TABLE serde_regex1(
+ key decimal,
+ value int)
+ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.RegexSerDe'
+WITH SERDEPROPERTIES (
+ "input.regex" = "([^ ]*) ([^ ]*)"
+)
+STORED AS TEXTFILE;
+
+LOAD DATA LOCAL INPATH "../data/files/kv7.txt" INTO TABLE serde_regex1;
+
+SELECT key, value FROM serde_regex1 ORDER BY key;
+
+DROP TABLE serde_regex1;
Modified: hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/smb_mapjoin_17.q
URL: http://svn.apache.org/viewvc/hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/smb_mapjoin_17.q?rev=1464915&r1=1464914&r2=1464915&view=diff
==============================================================================
--- hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/smb_mapjoin_17.q (original)
+++ hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/smb_mapjoin_17.q Fri Apr 5 10:34:08 2013
@@ -99,3 +99,27 @@ LEFT OUTER JOIN test_table5 e ON a.key =
LEFT OUTER JOIN test_table6 f ON a.key = f.key
LEFT OUTER JOIN test_table7 g ON a.key = g.key
LEFT OUTER JOIN test_table8 h ON a.key = h.key;
+
+-- outer join with max 16 aliases
+EXPLAIN
+SELECT a.*
+FROM test_table1 a
+LEFT OUTER JOIN test_table2 b ON a.key = b.key
+LEFT OUTER JOIN test_table3 c ON a.key = c.key
+LEFT OUTER JOIN test_table4 d ON a.key = d.key
+LEFT OUTER JOIN test_table5 e ON a.key = e.key
+LEFT OUTER JOIN test_table6 f ON a.key = f.key
+LEFT OUTER JOIN test_table7 g ON a.key = g.key
+LEFT OUTER JOIN test_table8 h ON a.key = h.key
+LEFT OUTER JOIN test_table4 i ON a.key = i.key
+LEFT OUTER JOIN test_table5 j ON a.key = j.key
+LEFT OUTER JOIN test_table6 k ON a.key = k.key
+LEFT OUTER JOIN test_table7 l ON a.key = l.key
+LEFT OUTER JOIN test_table8 m ON a.key = m.key
+LEFT OUTER JOIN test_table7 n ON a.key = n.key
+LEFT OUTER JOIN test_table8 o ON a.key = o.key
+LEFT OUTER JOIN test_table4 p ON a.key = p.key
+LEFT OUTER JOIN test_table5 q ON a.key = q.key
+LEFT OUTER JOIN test_table6 r ON a.key = r.key
+LEFT OUTER JOIN test_table7 s ON a.key = s.key
+LEFT OUTER JOIN test_table8 t ON a.key = t.key;
Modified: hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/union_remove_12.q
URL: http://svn.apache.org/viewvc/hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/union_remove_12.q?rev=1464915&r1=1464914&r2=1464915&view=diff
==============================================================================
--- hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/union_remove_12.q (original)
+++ hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/union_remove_12.q Fri Apr 5 10:34:08 2013
@@ -1,6 +1,7 @@
set hive.stats.autogather=false;
set hive.optimize.union.remove=true;
set hive.mapred.supports.subdirectories=true;
+set hive.auto.convert.join=true;
set hive.merge.mapfiles=true;
set hive.merge.mapredfiles=true;
@@ -31,7 +32,7 @@ SELECT * FROM
(
select key, 1 as values from inputTbl1
union all
-select /*+ mapjoin(a) */ a.key as key, b.val as values
+select a.key as key, b.val as values
FROM inputTbl1 a join inputTbl1 b on a.key=b.key
)c;
@@ -40,7 +41,7 @@ SELECT * FROM
(
select key, 1 as values from inputTbl1
union all
-select /*+ mapjoin(a) */ a.key as key, b.val as values
+select a.key as key, b.val as values
FROM inputTbl1 a join inputTbl1 b on a.key=b.key
)c;
Modified: hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/union_remove_13.q
URL: http://svn.apache.org/viewvc/hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/union_remove_13.q?rev=1464915&r1=1464914&r2=1464915&view=diff
==============================================================================
--- hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/union_remove_13.q (original)
+++ hive/branches/HIVE-4115/ql/src/test/queries/clientpositive/union_remove_13.q Fri Apr 5 10:34:08 2013
@@ -1,6 +1,7 @@
set hive.stats.autogather=false;
set hive.optimize.union.remove=true;
set hive.mapred.supports.subdirectories=true;
+set hive.auto.convert.join=true;
set hive.merge.mapfiles=true;
set hive.merge.mapredfiles=true;
@@ -31,7 +32,7 @@ SELECT * FROM
(
select key, count(1) as values from inputTbl1 group by key
union all
-select /*+ mapjoin(a) */ a.key as key, b.val as values
+select a.key as key, b.val as values
FROM inputTbl1 a join inputTbl1 b on a.key=b.key
)c;
@@ -40,7 +41,7 @@ SELECT * FROM
(
select key, count(1) as values from inputTbl1 group by key
union all
-select /*+ mapjoin(a) */ a.key as key, b.val as values
+select a.key as key, b.val as values
FROM inputTbl1 a join inputTbl1 b on a.key=b.key
)c;
Modified: hive/branches/HIVE-4115/ql/src/test/results/clientpositive/columnstats_tbllvl.q.out
URL: http://svn.apache.org/viewvc/hive/branches/HIVE-4115/ql/src/test/results/clientpositive/columnstats_tbllvl.q.out?rev=1464915&r1=1464914&r2=1464915&view=diff
==============================================================================
--- hive/branches/HIVE-4115/ql/src/test/results/clientpositive/columnstats_tbllvl.q.out (original)
+++ hive/branches/HIVE-4115/ql/src/test/results/clientpositive/columnstats_tbllvl.q.out Fri Apr 5 10:34:08 2013
@@ -264,3 +264,124 @@ POSTHOOK: query: analyze table UserVisit
POSTHOOK: type: QUERY
POSTHOOK: Input: default@uservisits_web_text_none
#### A masked pattern was here ####
+PREHOOK: query: CREATE TABLE empty_tab(
+ a int,
+ b double,
+ c string,
+ d boolean,
+ e binary)
+row format delimited fields terminated by '|' stored as textfile
+PREHOOK: type: CREATETABLE
+POSTHOOK: query: CREATE TABLE empty_tab(
+ a int,
+ b double,
+ c string,
+ d boolean,
+ e binary)
+row format delimited fields terminated by '|' stored as textfile
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: default@empty_tab
+PREHOOK: query: explain
+analyze table empty_tab compute statistics for columns a,b,c,d,e
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+analyze table empty_tab compute statistics for columns a,b,c,d,e
+POSTHOOK: type: QUERY
+ABSTRACT SYNTAX TREE:
+ (TOK_ANALYZE (TOK_TAB (TOK_TABNAME empty_tab)) (TOK_TABCOLNAME a b c d e))
+
+STAGE DEPENDENCIES:
+ Stage-0 is a root stage
+ Stage-1 is a root stage
+
+STAGE PLANS:
+ Stage: Stage-0
+ Map Reduce
+ Alias -> Map Operator Tree:
+ empty_tab
+ TableScan
+ alias: empty_tab
+ Select Operator
+ expressions:
+ expr: a
+ type: int
+ expr: b
+ type: double
+ expr: c
+ type: string
+ expr: d
+ type: boolean
+ expr: e
+ type: binary
+ outputColumnNames: a, b, c, d, e
+ Group By Operator
+ aggregations:
+ expr: compute_stats(a, 16)
+ expr: compute_stats(b, 16)
+ expr: compute_stats(c, 16)
+ expr: compute_stats(d, 16)
+ expr: compute_stats(e, 16)
+ bucketGroup: false
+ mode: hash
+ outputColumnNames: _col0, _col1, _col2, _col3, _col4
+ Reduce Output Operator
+ sort order:
+ tag: -1
+ value expressions:
+ expr: _col0
+ type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:string,numbitvectors:int>
+ expr: _col1
+ type: struct<columntype:string,min:double,max:double,countnulls:bigint,bitvector:string,numbitvectors:int>
+ expr: _col2
+ type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:string,numbitvectors:int>
+ expr: _col3
+ type: struct<columntype:string,counttrues:bigint,countfalses:bigint,countnulls:bigint>
+ expr: _col4
+ type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint>
+ Reduce Operator Tree:
+ Group By Operator
+ aggregations:
+ expr: compute_stats(VALUE._col0)
+ expr: compute_stats(VALUE._col1)
+ expr: compute_stats(VALUE._col2)
+ expr: compute_stats(VALUE._col3)
+ expr: compute_stats(VALUE._col4)
+ bucketGroup: false
+ mode: mergepartial
+ outputColumnNames: _col0, _col1, _col2, _col3, _col4
+ Select Operator
+ expressions:
+ expr: _col0
+ type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,numdistinctvalues:bigint>
+ expr: _col1
+ type: struct<columntype:string,min:double,max:double,countnulls:bigint,numdistinctvalues:bigint>
+ expr: _col2
+ type: struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint>
+ expr: _col3
+ type: struct<columntype:string,counttrues:bigint,countfalses:bigint,countnulls:bigint>
+ expr: _col4
+ type: struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint>
+ outputColumnNames: _col0, _col1, _col2, _col3, _col4
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+
+ Stage: Stage-1
+ Column Stats Work
+ Column Stats Desc:
+ Columns: a, b, c, d, e
+ Column Types: int, double, string, boolean, binary
+ Table: empty_tab
+
+
+PREHOOK: query: analyze table empty_tab compute statistics for columns a,b,c,d,e
+PREHOOK: type: QUERY
+PREHOOK: Input: default@empty_tab
+#### A masked pattern was here ####
+POSTHOOK: query: analyze table empty_tab compute statistics for columns a,b,c,d,e
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@empty_tab
+#### A masked pattern was here ####
Modified: hive/branches/HIVE-4115/ql/src/test/results/clientpositive/join_filters_overlap.q.out
URL: http://svn.apache.org/viewvc/hive/branches/HIVE-4115/ql/src/test/results/clientpositive/join_filters_overlap.q.out?rev=1464915&r1=1464914&r2=1464915&view=diff
==============================================================================
--- hive/branches/HIVE-4115/ql/src/test/results/clientpositive/join_filters_overlap.q.out (original)
+++ hive/branches/HIVE-4115/ql/src/test/results/clientpositive/join_filters_overlap.q.out Fri Apr 5 10:34:08 2013
@@ -697,7 +697,7 @@ STAGE PLANS:
Map-reduce partition columns:
expr: key
type: int
- tag: 3
+ tag: 2
value expressions:
expr: key
type: int
@@ -720,7 +720,7 @@ STAGE PLANS:
Map-reduce partition columns:
expr: key
type: int
- tag: 2
+ tag: 3
value expressions:
expr: key
type: int
@@ -776,16 +776,16 @@ STAGE PLANS:
Join Operator
condition map:
Outer Join 0 to 1
- Left Outer Join0 to 2
- Left Outer Join1 to 3
+ Left Outer Join1 to 2
+ Left Outer Join0 to 3
condition expressions:
0 {VALUE._col0} {VALUE._col1}
1 {VALUE._col0} {VALUE._col1}
2 {VALUE._col0} {VALUE._col1}
3 {VALUE._col0} {VALUE._col1}
filter mappings:
- 0 [1, 1, 2, 1]
- 1 [0, 1, 3, 1]
+ 0 [1, 1, 3, 1]
+ 1 [0, 1, 2, 1]
filter predicates:
0 {(VALUE._col1 = 50)} {(VALUE._col1 = 40)}
1 {(VALUE._col1 = 50)} {(VALUE._col1 = 60)}
@@ -803,14 +803,14 @@ STAGE PLANS:
type: int
expr: _col5
type: int
- expr: _col12
- type: int
- expr: _col13
- type: int
expr: _col8
type: int
expr: _col9
type: int
+ expr: _col12
+ type: int
+ expr: _col13
+ type: int
outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7
File Output Operator
compressed: false
Modified: hive/branches/HIVE-4115/ql/src/test/results/clientpositive/nonreserved_keywords_insert_into1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/HIVE-4115/ql/src/test/results/clientpositive/nonreserved_keywords_insert_into1.q.out?rev=1464915&r1=1464914&r2=1464915&view=diff
==============================================================================
--- hive/branches/HIVE-4115/ql/src/test/results/clientpositive/nonreserved_keywords_insert_into1.q.out (original)
+++ hive/branches/HIVE-4115/ql/src/test/results/clientpositive/nonreserved_keywords_insert_into1.q.out Fri Apr 5 10:34:08 2013
@@ -2,9 +2,9 @@ PREHOOK: query: DROP TABLE insert
PREHOOK: type: DROPTABLE
POSTHOOK: query: DROP TABLE insert
POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE insert (from INT, as STRING)
+PREHOOK: query: CREATE TABLE insert (key INT, as STRING)
PREHOOK: type: CREATETABLE
-POSTHOOK: query: CREATE TABLE insert (from INT, as STRING)
+POSTHOOK: query: CREATE TABLE insert (key INT, as STRING)
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: default@insert
PREHOOK: query: EXPLAIN INSERT INTO TABLE insert SELECT * FROM src LIMIT 100
@@ -84,7 +84,7 @@ POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
POSTHOOK: Output: default@insert
POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: insert.from EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
PREHOOK: query: SELECT SUM(HASH(hash)) FROM (
SELECT TRANSFORM(*) USING 'tr \t _' AS (hash) FROM insert
) t
@@ -98,14 +98,14 @@ POSTHOOK: type: QUERY
POSTHOOK: Input: default@insert
#### A masked pattern was here ####
POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: insert.from EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
10226524244
PREHOOK: query: EXPLAIN INSERT INTO TABLE insert SELECT * FROM src LIMIT 100
PREHOOK: type: QUERY
POSTHOOK: query: EXPLAIN INSERT INTO TABLE insert SELECT * FROM src LIMIT 100
POSTHOOK: type: QUERY
POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: insert.from EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
ABSTRACT SYNTAX TREE:
(TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_INSERT_INTO (TOK_TAB (TOK_TABNAME insert))) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_LIMIT 100)))
@@ -180,8 +180,8 @@ POSTHOOK: Input: default@src
POSTHOOK: Output: default@insert
POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: insert.from EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: insert.from EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
PREHOOK: query: SELECT SUM(HASH(sum)) FROM (
SELECT TRANSFORM(*) USING 'tr \t _' AS (sum) FROM insert
) t
@@ -196,8 +196,8 @@ POSTHOOK: Input: default@insert
#### A masked pattern was here ####
POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: insert.from EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: insert.from EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
20453048488
PREHOOK: query: SELECT COUNT(*) FROM insert
PREHOOK: type: QUERY
@@ -209,8 +209,8 @@ POSTHOOK: Input: default@insert
#### A masked pattern was here ####
POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: insert.from EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: insert.from EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
200
PREHOOK: query: EXPLAIN INSERT OVERWRITE TABLE insert SELECT * FROM src LIMIT 10
PREHOOK: type: QUERY
@@ -218,8 +218,8 @@ POSTHOOK: query: EXPLAIN INSERT OVERWRIT
POSTHOOK: type: QUERY
POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: insert.from EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: insert.from EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
ABSTRACT SYNTAX TREE:
(TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME insert))) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_LIMIT 10)))
@@ -295,9 +295,9 @@ POSTHOOK: Output: default@insert
POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: insert.from EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: insert.from EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: insert.from EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
PREHOOK: query: SELECT SUM(HASH(add)) FROM (
SELECT TRANSFORM(*) USING 'tr \t _' AS (add) FROM insert
) t
@@ -313,9 +313,9 @@ POSTHOOK: Input: default@insert
POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: insert.from EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: insert.from EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: insert.from EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-826625916
PREHOOK: query: DROP TABLE insert
PREHOOK: type: DROPTABLE
@@ -328,6 +328,6 @@ POSTHOOK: Output: default@insert
POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: insert.from EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: insert.from EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: insert.from EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
Modified: hive/branches/HIVE-4115/ql/src/test/results/clientpositive/serde_regex.q.out
URL: http://svn.apache.org/viewvc/hive/branches/HIVE-4115/ql/src/test/results/clientpositive/serde_regex.q.out?rev=1464915&r1=1464914&r2=1464915&view=diff
==============================================================================
--- hive/branches/HIVE-4115/ql/src/test/results/clientpositive/serde_regex.q.out (original)
+++ hive/branches/HIVE-4115/ql/src/test/results/clientpositive/serde_regex.q.out Fri Apr 5 10:34:08 2013
@@ -127,3 +127,124 @@ POSTHOOK: query: DROP TABLE serde_regex
POSTHOOK: type: DROPTABLE
POSTHOOK: Input: default@serde_regex
POSTHOOK: Output: default@serde_regex
+PREHOOK: query: EXPLAIN
+CREATE TABLE serde_regex1(
+ key decimal,
+ value int)
+ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.RegexSerDe'
+WITH SERDEPROPERTIES (
+ "input.regex" = "([^ ]*) ([^ ]*)"
+)
+STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
+POSTHOOK: query: EXPLAIN
+CREATE TABLE serde_regex1(
+ key decimal,
+ value int)
+ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.RegexSerDe'
+WITH SERDEPROPERTIES (
+ "input.regex" = "([^ ]*) ([^ ]*)"
+)
+STORED AS TEXTFILE
+POSTHOOK: type: CREATETABLE
+ABSTRACT SYNTAX TREE:
+ (TOK_CREATETABLE (TOK_TABNAME serde_regex1) TOK_LIKETABLE (TOK_TABCOLLIST (TOK_TABCOL key TOK_DECIMAL) (TOK_TABCOL value TOK_INT)) (TOK_TABLESERIALIZER (TOK_SERDENAME 'org.apache.hadoop.hive.serde2.RegexSerDe' (TOK_TABLEPROPERTIES (TOK_TABLEPROPLIST (TOK_TABLEPROPERTY "input.regex" "([^ ]*) ([^ ]*)"))))) TOK_TBLTEXTFILE)
+
+STAGE DEPENDENCIES:
+ Stage-0 is a root stage
+
+STAGE PLANS:
+ Stage: Stage-0
+ Create Table Operator:
+ Create Table
+ columns: key decimal, value int
+ if not exists: false
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ # buckets: -1
+ output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+ serde name: org.apache.hadoop.hive.serde2.RegexSerDe
+ serde properties:
+ input.regex ([^ ]*) ([^ ]*)
+ name: serde_regex1
+ isExternal: false
+
+
+PREHOOK: query: CREATE TABLE serde_regex1(
+ key decimal,
+ value int)
+ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.RegexSerDe'
+WITH SERDEPROPERTIES (
+ "input.regex" = "([^ ]*) ([^ ]*)"
+)
+STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
+POSTHOOK: query: CREATE TABLE serde_regex1(
+ key decimal,
+ value int)
+ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.RegexSerDe'
+WITH SERDEPROPERTIES (
+ "input.regex" = "([^ ]*) ([^ ]*)"
+)
+STORED AS TEXTFILE
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: default@serde_regex1
+PREHOOK: query: LOAD DATA LOCAL INPATH "../data/files/kv7.txt" INTO TABLE serde_regex1
+PREHOOK: type: LOAD
+PREHOOK: Output: default@serde_regex1
+POSTHOOK: query: LOAD DATA LOCAL INPATH "../data/files/kv7.txt" INTO TABLE serde_regex1
+POSTHOOK: type: LOAD
+POSTHOOK: Output: default@serde_regex1
+PREHOOK: query: SELECT key, value FROM serde_regex1 ORDER BY key
+PREHOOK: type: QUERY
+PREHOOK: Input: default@serde_regex1
+#### A masked pattern was here ####
+POSTHOOK: query: SELECT key, value FROM serde_regex1 ORDER BY key
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@serde_regex1
+#### A masked pattern was here ####
+-1234567890.123456789 -1234567890
+-4.4E+3 4400
+-1255.49 -1255
+-1.122 -11
+-1.12 -1
+-1.12 -1
+-0.333 0
+-0.33 0
+-0.3 0
+0 0
+0 0
+1E-99 0
+0.01 0
+0.02 0
+0.1 0
+0.2 0
+0.3 0
+0.33 0
+0.333 0
+0.9999999999999999999999999 1
+1 1
+1 1
+1.12 1
+1.122 1
+2 2
+2 2
+3.14 3
+3.14 3
+3.14 3
+3.14 4
+1E+1 10
+2E+1 20
+1E+2 100
+124 124
+125.2 125
+2E+2 200
+1234567890.12345678 1234567890
+1E+99 0
+PREHOOK: query: DROP TABLE serde_regex1
+PREHOOK: type: DROPTABLE
+PREHOOK: Input: default@serde_regex1
+PREHOOK: Output: default@serde_regex1
+POSTHOOK: query: DROP TABLE serde_regex1
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Input: default@serde_regex1
+POSTHOOK: Output: default@serde_regex1
Modified: hive/branches/HIVE-4115/ql/src/test/results/clientpositive/show_functions.q.out
URL: http://svn.apache.org/viewvc/hive/branches/HIVE-4115/ql/src/test/results/clientpositive/show_functions.q.out?rev=1464915&r1=1464914&r2=1464915&view=diff
==============================================================================
--- hive/branches/HIVE-4115/ql/src/test/results/clientpositive/show_functions.q.out (original)
+++ hive/branches/HIVE-4115/ql/src/test/results/clientpositive/show_functions.q.out Fri Apr 5 10:34:08 2013
@@ -47,12 +47,14 @@ count
covar_pop
covar_samp
create_union
+cume_dist
date_add
date_sub
datediff
day
dayofmonth
degrees
+dense_rank
div
e
elt
@@ -64,6 +66,7 @@ exp
explode
field
find_in_set
+first_value
floor
format_number
from_unixtime
@@ -83,7 +86,10 @@ isnotnull
isnull
java_method
json_tuple
+lag
+last_value
lcase
+lead
length
like
ln
@@ -105,10 +111,12 @@ named_struct
negative
ngrams
not
+ntile
nvl
or
parse_url
parse_url_tuple
+percent_rank
percentile
percentile_approx
pi
@@ -119,6 +127,7 @@ power
printf
radians
rand
+rank
reflect
reflect2
regexp
@@ -128,6 +137,7 @@ repeat
reverse
rlike
round
+row_number
rpad
rtrim
second
@@ -197,6 +207,7 @@ count
covar_pop
covar_samp
create_union
+cume_dist
PREHOOK: query: SHOW FUNCTIONS '.*e$'
PREHOOK: type: SHOWFUNCTIONS
POSTHOOK: query: SHOW FUNCTIONS '.*e$'
@@ -206,15 +217,18 @@ case
coalesce
e
explode
+first_value
from_unixtime
in_file
inline
json_tuple
+last_value
lcase
like
locate
minute
negative
+ntile
parse_url_tuple
percentile
positive
Modified: hive/branches/HIVE-4115/ql/src/test/results/clientpositive/smb_mapjoin_17.q.out
URL: http://svn.apache.org/viewvc/hive/branches/HIVE-4115/ql/src/test/results/clientpositive/smb_mapjoin_17.q.out?rev=1464915&r1=1464914&r2=1464915&view=diff
==============================================================================
--- hive/branches/HIVE-4115/ql/src/test/results/clientpositive/smb_mapjoin_17.q.out (original)
+++ hive/branches/HIVE-4115/ql/src/test/results/clientpositive/smb_mapjoin_17.q.out Fri Apr 5 10:34:08 2013
@@ -687,3 +687,231 @@ POSTHOOK: Lineage: test_table7.value SIM
POSTHOOK: Lineage: test_table8.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: test_table8.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
13126
+PREHOOK: query: -- outer join with max 16 aliases
+EXPLAIN
+SELECT a.*
+FROM test_table1 a
+LEFT OUTER JOIN test_table2 b ON a.key = b.key
+LEFT OUTER JOIN test_table3 c ON a.key = c.key
+LEFT OUTER JOIN test_table4 d ON a.key = d.key
+LEFT OUTER JOIN test_table5 e ON a.key = e.key
+LEFT OUTER JOIN test_table6 f ON a.key = f.key
+LEFT OUTER JOIN test_table7 g ON a.key = g.key
+LEFT OUTER JOIN test_table8 h ON a.key = h.key
+LEFT OUTER JOIN test_table4 i ON a.key = i.key
+LEFT OUTER JOIN test_table5 j ON a.key = j.key
+LEFT OUTER JOIN test_table6 k ON a.key = k.key
+LEFT OUTER JOIN test_table7 l ON a.key = l.key
+LEFT OUTER JOIN test_table8 m ON a.key = m.key
+LEFT OUTER JOIN test_table7 n ON a.key = n.key
+LEFT OUTER JOIN test_table8 o ON a.key = o.key
+LEFT OUTER JOIN test_table4 p ON a.key = p.key
+LEFT OUTER JOIN test_table5 q ON a.key = q.key
+LEFT OUTER JOIN test_table6 r ON a.key = r.key
+LEFT OUTER JOIN test_table7 s ON a.key = s.key
+LEFT OUTER JOIN test_table8 t ON a.key = t.key
+PREHOOK: type: QUERY
+POSTHOOK: query: -- outer join with max 16 aliases
+EXPLAIN
+SELECT a.*
+FROM test_table1 a
+LEFT OUTER JOIN test_table2 b ON a.key = b.key
+LEFT OUTER JOIN test_table3 c ON a.key = c.key
+LEFT OUTER JOIN test_table4 d ON a.key = d.key
+LEFT OUTER JOIN test_table5 e ON a.key = e.key
+LEFT OUTER JOIN test_table6 f ON a.key = f.key
+LEFT OUTER JOIN test_table7 g ON a.key = g.key
+LEFT OUTER JOIN test_table8 h ON a.key = h.key
+LEFT OUTER JOIN test_table4 i ON a.key = i.key
+LEFT OUTER JOIN test_table5 j ON a.key = j.key
+LEFT OUTER JOIN test_table6 k ON a.key = k.key
+LEFT OUTER JOIN test_table7 l ON a.key = l.key
+LEFT OUTER JOIN test_table8 m ON a.key = m.key
+LEFT OUTER JOIN test_table7 n ON a.key = n.key
+LEFT OUTER JOIN test_table8 o ON a.key = o.key
+LEFT OUTER JOIN test_table4 p ON a.key = p.key
+LEFT OUTER JOIN test_table5 q ON a.key = q.key
+LEFT OUTER JOIN test_table6 r ON a.key = r.key
+LEFT OUTER JOIN test_table7 s ON a.key = s.key
+LEFT OUTER JOIN test_table8 t ON a.key = t.key
+POSTHOOK: type: QUERY
+POSTHOOK: Lineage: test_table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: test_table1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: test_table2.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: test_table2.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: test_table3.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: test_table3.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: test_table4.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: test_table4.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: test_table5.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: test_table5.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: test_table6.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: test_table6.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: test_table7.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: test_table7.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: test_table8.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: test_table8.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
+ABSTRACT SYNTAX TREE:
+ (TOK_QUERY (TOK_FROM (TOK_LEFTOUTERJOIN (TOK_LEFTOUTERJOIN (TOK_LEFTOUTERJOIN (TOK_LEFTOUTERJOIN (TOK_LEFTOUTERJOIN (TOK_LEFTOUTERJOIN (TOK_LEFTOUTERJOIN (TOK_LEFTOUTERJOIN (TOK_LEFTOUTERJOIN (TOK_LEFTOUTERJOIN (TOK_LEFTOUTERJOIN (TOK_LEFTOUTERJOIN (TOK_LEFTOUTERJOIN (TOK_LEFTOUTERJOIN (TOK_LEFTOUTERJOIN (TOK_LEFTOUTERJOIN (TOK_LEFTOUTERJOIN (TOK_LEFTOUTERJOIN (TOK_LEFTOUTERJOIN (TOK_TABREF (TOK_TABNAME test_table1) a) (TOK_TABREF (TOK_TABNAME test_table2) b) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL b) key))) (TOK_TABREF (TOK_TABNAME test_table3) c) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL c) key))) (TOK_TABREF (TOK_TABNAME test_table4) d) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL d) key))) (TOK_TABREF (TOK_TABNAME test_table5) e) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL e) key))) (TOK_TABREF (TOK_TABNAME test_table6) f) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL f) key))) (TOK_TABREF (TOK_TABNAME test_table7) g) (=
(. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL g) key))) (TOK_TABREF (TOK_TABNAME test_table8) h) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL h) key))) (TOK_TABREF (TOK_TABNAME test_table4) i) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL i) key))) (TOK_TABREF (TOK_TABNAME test_table5) j) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL j) key))) (TOK_TABREF (TOK_TABNAME test_table6) k) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL k) key))) (TOK_TABREF (TOK_TABNAME test_table7) l) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL l) key))) (TOK_TABREF (TOK_TABNAME test_table8) m) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL m) key))) (TOK_TABREF (TOK_TABNAME test_table7) n) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL n) key))) (TOK_TABREF (TOK_TABNAME test_table8) o) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL o) key))) (TOK_TABREF (TOK_TABNAME test_table4) p) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL p)
key))) (TOK_TABREF (TOK_TABNAME test_table5) q) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL q) key))) (TOK_TABREF (TOK_TABNAME test_table6) r) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL r) key))) (TOK_TABREF (TOK_TABNAME test_table7) s) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL s) key))) (TOK_TABREF (TOK_TABNAME test_table8) t) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL t) key)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME a))))))
+
+STAGE DEPENDENCIES:
+ Stage-1 is a root stage
+ Stage-0 is a root stage
+
+STAGE PLANS:
+ Stage: Stage-1
+ Map Reduce
+ Alias -> Map Operator Tree:
+ a
+ TableScan
+ alias: a
+ Sorted Merge Bucket Map Join Operator
+ condition map:
+ Left Outer Join0 to 1
+ Left Outer Join0 to 2
+ Left Outer Join0 to 3
+ Left Outer Join0 to 4
+ Left Outer Join0 to 5
+ Left Outer Join0 to 6
+ Left Outer Join0 to 7
+ Left Outer Join0 to 8
+ Left Outer Join0 to 9
+ Left Outer Join0 to 10
+ Left Outer Join0 to 11
+ Left Outer Join0 to 12
+ Left Outer Join0 to 13
+ Left Outer Join0 to 14
+ Left Outer Join0 to 15
+ condition expressions:
+ 0 {key} {value}
+ 1
+ 2
+ 3
+ 4
+ 5
+ 6
+ 7
+ 8
+ 9
+ 10
+ 11
+ 12
+ 13
+ 14
+ 15
+ handleSkewJoin: false
+ keys:
+ 0 [Column[key]]
+ 1 [Column[key]]
+ 2 [Column[key]]
+ 3 [Column[key]]
+ 4 [Column[key]]
+ 5 [Column[key]]
+ 6 [Column[key]]
+ 7 [Column[key]]
+ 8 [Column[key]]
+ 9 [Column[key]]
+ 10 [Column[key]]
+ 11 [Column[key]]
+ 12 [Column[key]]
+ 13 [Column[key]]
+ 14 [Column[key]]
+ 15 [Column[key]]
+ outputColumnNames: _col0, _col1
+ Position of Big Table: 0
+ Reduce Output Operator
+ key expressions:
+ expr: _col0
+ type: int
+ sort order: +
+ Map-reduce partition columns:
+ expr: _col0
+ type: int
+ tag: 0
+ value expressions:
+ expr: _col0
+ type: int
+ expr: _col1
+ type: string
+ q
+ TableScan
+ alias: q
+ Reduce Output Operator
+ key expressions:
+ expr: key
+ type: int
+ sort order: +
+ Map-reduce partition columns:
+ expr: key
+ type: int
+ tag: 1
+ r
+ TableScan
+ alias: r
+ Reduce Output Operator
+ key expressions:
+ expr: key
+ type: int
+ sort order: +
+ Map-reduce partition columns:
+ expr: key
+ type: int
+ tag: 2
+ s
+ TableScan
+ alias: s
+ Reduce Output Operator
+ key expressions:
+ expr: key
+ type: int
+ sort order: +
+ Map-reduce partition columns:
+ expr: key
+ type: int
+ tag: 3
+ t
+ TableScan
+ alias: t
+ Reduce Output Operator
+ key expressions:
+ expr: key
+ type: int
+ sort order: +
+ Map-reduce partition columns:
+ expr: key
+ type: int
+ tag: 4
+ Reduce Operator Tree:
+ Join Operator
+ condition map:
+ Left Outer Join0 to 1
+ Left Outer Join0 to 2
+ Left Outer Join0 to 3
+ Left Outer Join0 to 4
+ condition expressions:
+ 0 {VALUE._col24} {VALUE._col25}
+ 1
+ 2
+ 3
+ 4
+ handleSkewJoin: false
+ outputColumnNames: _col24, _col25
+ Select Operator
+ expressions:
+ expr: _col24
+ type: int
+ expr: _col25
+ type: string
+ outputColumnNames: _col0, _col1
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+
+ Stage: Stage-0
+ Fetch Operator
+ limit: -1
+
+
Modified: hive/branches/HIVE-4115/ql/src/test/results/clientpositive/udf_round.q.out
URL: http://svn.apache.org/viewvc/hive/branches/HIVE-4115/ql/src/test/results/clientpositive/udf_round.q.out?rev=1464915&r1=1464914&r2=1464915&view=diff
==============================================================================
--- hive/branches/HIVE-4115/ql/src/test/results/clientpositive/udf_round.q.out (original)
+++ hive/branches/HIVE-4115/ql/src/test/results/clientpositive/udf_round.q.out Fri Apr 5 10:34:08 2013
@@ -40,7 +40,7 @@ FROM src LIMIT 1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
#### A masked pattern was here ####
-55555.0 55555.0 55555.0 55555.0 55555.0 55560.0 55600.0 56000.0 60000.0 100000.0 0.0 0.0 0.0
+55555 55555.0 55555.0 55555.0 55555.0 55560.0 55600.0 56000.0 60000.0 100000.0 0.0 0.0 0.0
PREHOOK: query: SELECT
round(125.315), round(125.315, 0),
round(125.315, 1), round(125.315, 2), round(125.315, 3), round(125.315, 4),
Modified: hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_12.q.out
URL: http://svn.apache.org/viewvc/hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_12.q.out?rev=1464915&r1=1464914&r2=1464915&view=diff
==============================================================================
--- hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_12.q.out (original)
+++ hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_12.q.out Fri Apr 5 10:34:08 2013
@@ -46,7 +46,7 @@ SELECT * FROM
(
select key, 1 as values from inputTbl1
union all
-select /*+ mapjoin(a) */ a.key as key, b.val as values
+select a.key as key, b.val as values
FROM inputTbl1 a join inputTbl1 b on a.key=b.key
)c
PREHOOK: type: QUERY
@@ -56,27 +56,26 @@ SELECT * FROM
(
select key, 1 as values from inputTbl1
union all
-select /*+ mapjoin(a) */ a.key as key, b.val as values
+select a.key as key, b.val as values
FROM inputTbl1 a join inputTbl1 b on a.key=b.key
)c
POSTHOOK: type: QUERY
ABSTRACT SYNTAX TREE:
- (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_UNION (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME inputTbl1))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR 1 values)))) (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_TABREF (TOK_TABNAME inputTbl1) a) (TOK_TABREF (TOK_TABNAME inputTbl1) b) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL b) key)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_HINTLIST (TOK_HINT TOK_MAPJOIN (TOK_HINTARGLIST a))) (TOK_SELEXPR (. (TOK_TABLE_OR_COL a) key) key) (TOK_SELEXPR (. (TOK_TABLE_OR_COL b) val) values))))) c)) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME outputTbl1))) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF))))
+ (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_UNION (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME inputTbl1))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR 1 values)))) (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_TABREF (TOK_TABNAME inputTbl1) a) (TOK_TABREF (TOK_TABNAME inputTbl1) b) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL b) key)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL a) key) key) (TOK_SELEXPR (. (TOK_TABLE_OR_COL b) val) values))))) c)) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME outputTbl1))) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF))))
STAGE DEPENDENCIES:
- Stage-9 is a root stage
- Stage-7 depends on stages: Stage-2, Stage-9 , consists of Stage-4, Stage-3, Stage-5
- Stage-4
- Stage-0 depends on stages: Stage-4, Stage-3, Stage-6
+ Stage-8 is a root stage
+ Stage-6 depends on stages: Stage-8, Stage-9 , consists of Stage-3, Stage-2, Stage-4
Stage-3
- Stage-5
- Stage-6 depends on stages: Stage-5
+ Stage-0 depends on stages: Stage-3, Stage-2, Stage-5
+ Stage-2
+ Stage-4
+ Stage-5 depends on stages: Stage-4
Stage-10 is a root stage
- Stage-1 depends on stages: Stage-10
- Stage-2 depends on stages: Stage-1
+ Stage-9 depends on stages: Stage-10
STAGE PLANS:
- Stage: Stage-9
+ Stage: Stage-8
Map Reduce
Alias -> Map Operator Tree:
null-subquery1:c-subquery1:inputtbl1
@@ -86,43 +85,22 @@ STAGE PLANS:
expressions:
expr: key
type: string
- expr: 1
- type: int
+ expr: UDFToLong(UDFToString(1))
+ type: bigint
outputColumnNames: _col0, _col1
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: UDFToString(_col1)
- type: string
- outputColumnNames: _col0, _col1
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: _col1
- type: string
- outputColumnNames: _col0, _col1
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: UDFToLong(_col1)
- type: bigint
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- GlobalTableId: 1
- table:
- input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
- serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
- name: default.outputtbl1
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ table:
+ input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
+ name: default.outputtbl1
- Stage: Stage-7
+ Stage: Stage-6
Conditional Operator
- Stage: Stage-4
+ Stage: Stage-3
Move Operator
files:
hdfs directory: true
@@ -138,13 +116,13 @@ STAGE PLANS:
serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
name: default.outputtbl1
- Stage: Stage-3
+ Stage: Stage-2
Block level merge
- Stage: Stage-5
+ Stage: Stage-4
Block level merge
- Stage: Stage-6
+ Stage: Stage-5
Move Operator
files:
hdfs directory: true
@@ -170,7 +148,7 @@ STAGE PLANS:
1 [Column[key]]
Position of Big Table: 1
- Stage: Stage-1
+ Stage: Stage-9
Map Reduce
Alias -> Map Operator Tree:
null-subquery2:c-subquery2:b
@@ -188,55 +166,23 @@ STAGE PLANS:
1 [Column[key]]
outputColumnNames: _col0, _col5
Position of Big Table: 1
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.SequenceFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
- Local Work:
- Map Reduce Local Work
-
- Stage: Stage-2
- Map Reduce
- Alias -> Map Operator Tree:
-#### A masked pattern was here ####
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: _col5
- type: string
- outputColumnNames: _col0, _col5
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: _col5
- type: string
- outputColumnNames: _col0, _col1
Select Operator
expressions:
expr: _col0
type: string
- expr: _col1
- type: string
+ expr: UDFToLong(_col5)
+ type: bigint
outputColumnNames: _col0, _col1
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: UDFToLong(_col1)
- type: bigint
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- GlobalTableId: 1
- table:
- input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
- serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
- name: default.outputtbl1
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ table:
+ input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
+ name: default.outputtbl1
+ Local Work:
+ Map Reduce Local Work
PREHOOK: query: insert overwrite table outputTbl1
@@ -244,7 +190,7 @@ SELECT * FROM
(
select key, 1 as values from inputTbl1
union all
-select /*+ mapjoin(a) */ a.key as key, b.val as values
+select a.key as key, b.val as values
FROM inputTbl1 a join inputTbl1 b on a.key=b.key
)c
PREHOOK: type: QUERY
@@ -255,7 +201,7 @@ SELECT * FROM
(
select key, 1 as values from inputTbl1
union all
-select /*+ mapjoin(a) */ a.key as key, b.val as values
+select a.key as key, b.val as values
FROM inputTbl1 a join inputTbl1 b on a.key=b.key
)c
POSTHOOK: type: QUERY
Modified: hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_13.q.out
URL: http://svn.apache.org/viewvc/hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_13.q.out?rev=1464915&r1=1464914&r2=1464915&view=diff
==============================================================================
--- hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_13.q.out (original)
+++ hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_13.q.out Fri Apr 5 10:34:08 2013
@@ -46,7 +46,7 @@ SELECT * FROM
(
select key, count(1) as values from inputTbl1 group by key
union all
-select /*+ mapjoin(a) */ a.key as key, b.val as values
+select a.key as key, b.val as values
FROM inputTbl1 a join inputTbl1 b on a.key=b.key
)c
PREHOOK: type: QUERY
@@ -56,27 +56,26 @@ SELECT * FROM
(
select key, count(1) as values from inputTbl1 group by key
union all
-select /*+ mapjoin(a) */ a.key as key, b.val as values
+select a.key as key, b.val as values
FROM inputTbl1 a join inputTbl1 b on a.key=b.key
)c
POSTHOOK: type: QUERY
ABSTRACT SYNTAX TREE:
- (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_UNION (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME inputTbl1))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_FUNCTION count 1) values)) (TOK_GROUPBY (TOK_TABLE_OR_COL key)))) (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_TABREF (TOK_TABNAME inputTbl1) a) (TOK_TABREF (TOK_TABNAME inputTbl1) b) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL b) key)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_HINTLIST (TOK_HINT TOK_MAPJOIN (TOK_HINTARGLIST a))) (TOK_SELEXPR (. (TOK_TABLE_OR_COL a) key) key) (TOK_SELEXPR (. (TOK_TABLE_OR_COL b) val) values))))) c)) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME outputTbl1))) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF))))
+ (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_UNION (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME inputTbl1))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_FUNCTION count 1) values)) (TOK_GROUPBY (TOK_TABLE_OR_COL key)))) (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_TABREF (TOK_TABNAME inputTbl1) a) (TOK_TABREF (TOK_TABNAME inputTbl1) b) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL b) key)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL a) key) key) (TOK_SELEXPR (. (TOK_TABLE_OR_COL b) val) values))))) c)) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME outputTbl1))) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF))))
STAGE DEPENDENCIES:
- Stage-9 is a root stage
- Stage-7 depends on stages: Stage-2, Stage-9 , consists of Stage-4, Stage-3, Stage-5
- Stage-4
- Stage-0 depends on stages: Stage-4, Stage-3, Stage-6
+ Stage-8 is a root stage
+ Stage-6 depends on stages: Stage-8, Stage-9 , consists of Stage-3, Stage-2, Stage-4
Stage-3
- Stage-5
- Stage-6 depends on stages: Stage-5
+ Stage-0 depends on stages: Stage-3, Stage-2, Stage-5
+ Stage-2
+ Stage-4
+ Stage-5 depends on stages: Stage-4
Stage-10 is a root stage
- Stage-1 depends on stages: Stage-10
- Stage-2 depends on stages: Stage-1
+ Stage-9 depends on stages: Stage-10
STAGE PLANS:
- Stage: Stage-9
+ Stage: Stage-8
Map Reduce
Alias -> Map Operator Tree:
null-subquery1:c-subquery1:inputtbl1
@@ -122,43 +121,22 @@ STAGE PLANS:
expressions:
expr: _col0
type: string
- expr: _col1
+ expr: UDFToLong(UDFToString(_col1))
type: bigint
outputColumnNames: _col0, _col1
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: UDFToString(_col1)
- type: string
- outputColumnNames: _col0, _col1
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: _col1
- type: string
- outputColumnNames: _col0, _col1
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: UDFToLong(_col1)
- type: bigint
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- GlobalTableId: 1
- table:
- input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
- serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
- name: default.outputtbl1
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ table:
+ input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
+ name: default.outputtbl1
- Stage: Stage-7
+ Stage: Stage-6
Conditional Operator
- Stage: Stage-4
+ Stage: Stage-3
Move Operator
files:
hdfs directory: true
@@ -174,13 +152,13 @@ STAGE PLANS:
serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
name: default.outputtbl1
- Stage: Stage-3
+ Stage: Stage-2
Block level merge
- Stage: Stage-5
+ Stage: Stage-4
Block level merge
- Stage: Stage-6
+ Stage: Stage-5
Move Operator
files:
hdfs directory: true
@@ -206,7 +184,7 @@ STAGE PLANS:
1 [Column[key]]
Position of Big Table: 1
- Stage: Stage-1
+ Stage: Stage-9
Map Reduce
Alias -> Map Operator Tree:
null-subquery2:c-subquery2:b
@@ -224,55 +202,23 @@ STAGE PLANS:
1 [Column[key]]
outputColumnNames: _col0, _col5
Position of Big Table: 1
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.SequenceFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
- Local Work:
- Map Reduce Local Work
-
- Stage: Stage-2
- Map Reduce
- Alias -> Map Operator Tree:
-#### A masked pattern was here ####
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: _col5
- type: string
- outputColumnNames: _col0, _col5
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: _col5
- type: string
- outputColumnNames: _col0, _col1
Select Operator
expressions:
expr: _col0
type: string
- expr: _col1
- type: string
+ expr: UDFToLong(_col5)
+ type: bigint
outputColumnNames: _col0, _col1
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: UDFToLong(_col1)
- type: bigint
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- GlobalTableId: 1
- table:
- input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
- serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
- name: default.outputtbl1
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ table:
+ input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
+ name: default.outputtbl1
+ Local Work:
+ Map Reduce Local Work
PREHOOK: query: insert overwrite table outputTbl1
@@ -280,7 +226,7 @@ SELECT * FROM
(
select key, count(1) as values from inputTbl1 group by key
union all
-select /*+ mapjoin(a) */ a.key as key, b.val as values
+select a.key as key, b.val as values
FROM inputTbl1 a join inputTbl1 b on a.key=b.key
)c
PREHOOK: type: QUERY
@@ -291,7 +237,7 @@ SELECT * FROM
(
select key, count(1) as values from inputTbl1 group by key
union all
-select /*+ mapjoin(a) */ a.key as key, b.val as values
+select a.key as key, b.val as values
FROM inputTbl1 a join inputTbl1 b on a.key=b.key
)c
POSTHOOK: type: QUERY
Modified: hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_14.q.out
URL: http://svn.apache.org/viewvc/hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_14.q.out?rev=1464915&r1=1464914&r2=1464915&view=diff
==============================================================================
--- hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_14.q.out (original)
+++ hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_14.q.out Fri Apr 5 10:34:08 2013
@@ -67,18 +67,14 @@ ABSTRACT SYNTAX TREE:
STAGE DEPENDENCIES:
Stage-8 is a root stage
- Stage-6 depends on stages: Stage-1, Stage-8, Stage-9, Stage-10 , consists of Stage-3, Stage-2, Stage-4
+ Stage-6 depends on stages: Stage-8, Stage-9 , consists of Stage-3, Stage-2, Stage-4
Stage-3
Stage-0 depends on stages: Stage-3, Stage-2, Stage-5
Stage-2
Stage-4
Stage-5 depends on stages: Stage-4
- Stage-11 is a root stage , consists of Stage-12, Stage-13, Stage-1
- Stage-12 has a backup stage: Stage-1
- Stage-9 depends on stages: Stage-12
- Stage-13 has a backup stage: Stage-1
- Stage-10 depends on stages: Stage-13
- Stage-1
+ Stage-10 is a root stage
+ Stage-9 depends on stages: Stage-10
STAGE PLANS:
Stage: Stage-8
@@ -91,7 +87,7 @@ STAGE PLANS:
expressions:
expr: key
type: string
- expr: UDFToLong(1)
+ expr: UDFToLong(UDFToString(1))
type: bigint
outputColumnNames: _col0, _col1
File Output Operator
@@ -134,66 +130,7 @@ STAGE PLANS:
hdfs directory: true
#### A masked pattern was here ####
- Stage: Stage-11
- Conditional Operator
-
- Stage: Stage-12
- Map Reduce Local Work
- Alias -> Map Local Tables:
- null-subquery2:c-subquery2:b
- Fetch Operator
- limit: -1
- Alias -> Map Local Operator Tree:
- null-subquery2:c-subquery2:b
- TableScan
- alias: b
- HashTable Sink Operator
- condition expressions:
- 0 {key}
- 1 {val}
- handleSkewJoin: false
- keys:
- 0 [Column[key]]
- 1 [Column[key]]
- Position of Big Table: 0
-
- Stage: Stage-9
- Map Reduce
- Alias -> Map Operator Tree:
- null-subquery2:c-subquery2:a
- TableScan
- alias: a
- Map Join Operator
- condition map:
- Inner Join 0 to 1
- condition expressions:
- 0 {key}
- 1 {val}
- handleSkewJoin: false
- keys:
- 0 [Column[key]]
- 1 [Column[key]]
- outputColumnNames: _col0, _col5
- Position of Big Table: 0
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: UDFToLong(_col5)
- type: bigint
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- GlobalTableId: 1
- table:
- input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
- serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
- name: default.outputtbl1
- Local Work:
- Map Reduce Local Work
-
- Stage: Stage-13
+ Stage: Stage-10
Map Reduce Local Work
Alias -> Map Local Tables:
null-subquery2:c-subquery2:a
@@ -213,7 +150,7 @@ STAGE PLANS:
1 [Column[key]]
Position of Big Table: 1
- Stage: Stage-10
+ Stage: Stage-9
Map Reduce
Alias -> Map Operator Tree:
null-subquery2:c-subquery2:b
@@ -249,64 +186,6 @@ STAGE PLANS:
Local Work:
Map Reduce Local Work
- Stage: Stage-1
- Map Reduce
- Alias -> Map Operator Tree:
- null-subquery2:c-subquery2:a
- TableScan
- alias: a
- Reduce Output Operator
- key expressions:
- expr: key
- type: string
- sort order: +
- Map-reduce partition columns:
- expr: key
- type: string
- tag: 0
- value expressions:
- expr: key
- type: string
- null-subquery2:c-subquery2:b
- TableScan
- alias: b
- Reduce Output Operator
- key expressions:
- expr: key
- type: string
- sort order: +
- Map-reduce partition columns:
- expr: key
- type: string
- tag: 1
- value expressions:
- expr: val
- type: string
- Reduce Operator Tree:
- Join Operator
- condition map:
- Inner Join 0 to 1
- condition expressions:
- 0 {VALUE._col0}
- 1 {VALUE._col1}
- handleSkewJoin: false
- outputColumnNames: _col0, _col5
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: UDFToLong(_col5)
- type: bigint
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- GlobalTableId: 1
- table:
- input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
- serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
- name: default.outputtbl1
-
PREHOOK: query: insert overwrite table outputTbl1
SELECT * FROM
Modified: hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_2.q.out?rev=1464915&r1=1464914&r2=1464915&view=diff
==============================================================================
--- hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_2.q.out (original)
+++ hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_2.q.out Fri Apr 5 10:34:08 2013
@@ -78,8 +78,8 @@ STAGE PLANS:
expressions:
expr: key
type: string
- expr: 2
- type: int
+ expr: UDFToLong(2)
+ type: bigint
outputColumnNames: _col0, _col1
File Output Operator
compressed: false
@@ -168,8 +168,8 @@ STAGE PLANS:
expressions:
expr: key
type: string
- expr: 1
- type: int
+ expr: UDFToLong(1)
+ type: bigint
outputColumnNames: _col0, _col1
File Output Operator
compressed: false
Modified: hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_22.q.out
URL: http://svn.apache.org/viewvc/hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_22.q.out?rev=1464915&r1=1464914&r2=1464915&view=diff
==============================================================================
--- hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_22.q.out (original)
+++ hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_22.q.out Fri Apr 5 10:34:08 2013
@@ -109,24 +109,17 @@ STAGE PLANS:
type: string
expr: _col1
type: bigint
- outputColumnNames: _col0, _col1
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: _col1
- type: bigint
- expr: _col1
- type: bigint
- outputColumnNames: _col0, _col1, _col2
- File Output Operator
- compressed: false
- GlobalTableId: 1
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.outputtbl1
+ expr: _col1
+ type: bigint
+ outputColumnNames: _col0, _col1, _col2
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.outputtbl1
Stage: Stage-0
Move Operator
@@ -186,24 +179,17 @@ STAGE PLANS:
type: string
expr: _col1
type: bigint
- outputColumnNames: _col0, _col1
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: _col1
- type: bigint
- expr: _col1
- type: bigint
- outputColumnNames: _col0, _col1, _col2
- File Output Operator
- compressed: false
- GlobalTableId: 1
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.outputtbl1
+ expr: _col1
+ type: bigint
+ outputColumnNames: _col0, _col1, _col2
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.outputtbl1
PREHOOK: query: insert overwrite table outputTbl1
@@ -359,35 +345,19 @@ STAGE PLANS:
expressions:
expr: _col0
type: string
- expr: _col1
+ expr: UDFToLong(concat(_col1, _col1))
type: bigint
- outputColumnNames: _col0, _col1
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: concat(_col1, _col1)
- type: string
- expr: concat(_col1, _col1)
- type: string
- outputColumnNames: _col0, _col1, _col2
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: UDFToLong(_col1)
- type: bigint
- expr: UDFToLong(_col2)
- type: bigint
- outputColumnNames: _col0, _col1, _col2
- File Output Operator
- compressed: false
- GlobalTableId: 1
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.outputtbl1
+ expr: UDFToLong(concat(_col1, _col1))
+ type: bigint
+ outputColumnNames: _col0, _col1, _col2
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.outputtbl1
Stage: Stage-0
Move Operator
@@ -445,35 +415,19 @@ STAGE PLANS:
expressions:
expr: _col0
type: string
- expr: _col1
+ expr: UDFToLong(concat(_col1, _col1))
type: bigint
- outputColumnNames: _col0, _col1
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: concat(_col1, _col1)
- type: string
- expr: concat(_col1, _col1)
- type: string
- outputColumnNames: _col0, _col1, _col2
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: UDFToLong(_col1)
- type: bigint
- expr: UDFToLong(_col2)
- type: bigint
- outputColumnNames: _col0, _col1, _col2
- File Output Operator
- compressed: false
- GlobalTableId: 1
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.outputtbl1
+ expr: UDFToLong(concat(_col1, _col1))
+ type: bigint
+ outputColumnNames: _col0, _col1, _col2
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.outputtbl1
PREHOOK: query: insert overwrite table outputTbl1
Modified: hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_23.q.out
URL: http://svn.apache.org/viewvc/hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_23.q.out?rev=1464915&r1=1464914&r2=1464915&view=diff
==============================================================================
--- hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_23.q.out (original)
+++ hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_23.q.out Fri Apr 5 10:34:08 2013
@@ -115,21 +115,14 @@ STAGE PLANS:
expr: _col1
type: bigint
outputColumnNames: _col0, _col1
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: _col1
- type: bigint
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- GlobalTableId: 1
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.outputtbl1
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.outputtbl1
Stage: Stage-0
Move Operator
@@ -185,26 +178,21 @@ STAGE PLANS:
expr: _col0
type: string
outputColumnNames: _col0
- Select Operator
- expressions:
+ Group By Operator
+ aggregations:
+ expr: count(1)
+ bucketGroup: false
+ keys:
expr: _col0
type: string
- outputColumnNames: _col0
- Group By Operator
- aggregations:
- expr: count(1)
- bucketGroup: false
- keys:
- expr: _col0
- type: string
- mode: hash
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- GlobalTableId: 0
- table:
- input format: org.apache.hadoop.mapred.SequenceFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ mode: hash
+ outputColumnNames: _col0, _col1
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
Stage: Stage-3
Map Reduce
@@ -239,21 +227,14 @@ STAGE PLANS:
expr: _col1
type: bigint
outputColumnNames: _col0, _col1
- Select Operator
- expressions:
- expr: _col0
- type: string
- expr: _col1
- type: bigint
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- GlobalTableId: 1
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.outputtbl1
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.outputtbl1
PREHOOK: query: insert overwrite table outputTbl1
Modified: hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_24.q.out
URL: http://svn.apache.org/viewvc/hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_24.q.out?rev=1464915&r1=1464914&r2=1464915&view=diff
==============================================================================
--- hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_24.q.out (original)
+++ hive/branches/HIVE-4115/ql/src/test/results/clientpositive/union_remove_24.q.out Fri Apr 5 10:34:08 2013
@@ -103,33 +103,19 @@ STAGE PLANS:
outputColumnNames: _col0, _col1
Select Operator
expressions:
- expr: UDFToLong(_col0)
- type: bigint
+ expr: UDFToDouble(UDFToLong(_col0))
+ type: double
expr: _col1
type: bigint
outputColumnNames: _col0, _col1
- Select Operator
- expressions:
- expr: UDFToDouble(_col0)
- type: double
- expr: _col1
- type: bigint
- outputColumnNames: _col0, _col1
- Select Operator
- expressions:
- expr: _col0
- type: double
- expr: _col1
- type: bigint
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- GlobalTableId: 1
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.outputtbl1
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.outputtbl1
Stage: Stage-0
Move Operator
@@ -190,21 +176,14 @@ STAGE PLANS:
expr: _col1
type: bigint
outputColumnNames: _col0, _col1
- Select Operator
- expressions:
- expr: _col0
- type: double
- expr: _col1
- type: bigint
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- GlobalTableId: 1
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.outputtbl1
+ File Output Operator
+ compressed: false
+ GlobalTableId: 1
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.outputtbl1
PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1