You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by jc...@apache.org on 2018/01/13 02:14:11 UTC
[18/23] hive git commit: HIVE-18416: Initial support for TABLE
function (Jesus Camacho Rodriguez, reviewed by Ashutosh Chauhan)
http://git-wip-us.apache.org/repos/asf/hive/blob/7e64114d/ql/src/test/results/clientpositive/join43.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/join43.q.out b/ql/src/test/results/clientpositive/join43.q.out
index fac8a8c..dbb3e33 100644
--- a/ql/src/test/results/clientpositive/join43.q.out
+++ b/ql/src/test/results/clientpositive/join43.q.out
@@ -8,44 +8,52 @@ POSTHOOK: Output: database:default
POSTHOOK: Output: default@purchase_history
PREHOOK: query: insert into purchase_history values ('1', 'Belt', 20.00, 21)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@purchase_history
POSTHOOK: query: insert into purchase_history values ('1', 'Belt', 20.00, 21)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@purchase_history
-POSTHOOK: Lineage: purchase_history.price EXPRESSION [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
-POSTHOOK: Lineage: purchase_history.product SIMPLE [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: purchase_history.s SIMPLE [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: purchase_history.time EXPRESSION [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col4, type:string, comment:), ]
+POSTHOOK: Lineage: purchase_history.price SCRIPT []
+POSTHOOK: Lineage: purchase_history.product SCRIPT []
+POSTHOOK: Lineage: purchase_history.s SCRIPT []
+POSTHOOK: Lineage: purchase_history.time SCRIPT []
PREHOOK: query: insert into purchase_history values ('1', 'Socks', 3.50, 31)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@purchase_history
POSTHOOK: query: insert into purchase_history values ('1', 'Socks', 3.50, 31)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@purchase_history
-POSTHOOK: Lineage: purchase_history.price EXPRESSION [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
-POSTHOOK: Lineage: purchase_history.product SIMPLE [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: purchase_history.s SIMPLE [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: purchase_history.time EXPRESSION [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col4, type:string, comment:), ]
+POSTHOOK: Lineage: purchase_history.price SCRIPT []
+POSTHOOK: Lineage: purchase_history.product SCRIPT []
+POSTHOOK: Lineage: purchase_history.s SCRIPT []
+POSTHOOK: Lineage: purchase_history.time SCRIPT []
PREHOOK: query: insert into purchase_history values ('3', 'Belt', 20.00, 51)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@purchase_history
POSTHOOK: query: insert into purchase_history values ('3', 'Belt', 20.00, 51)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@purchase_history
-POSTHOOK: Lineage: purchase_history.price EXPRESSION [(values__tmp__table__3)values__tmp__table__3.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
-POSTHOOK: Lineage: purchase_history.product SIMPLE [(values__tmp__table__3)values__tmp__table__3.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: purchase_history.s SIMPLE [(values__tmp__table__3)values__tmp__table__3.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: purchase_history.time EXPRESSION [(values__tmp__table__3)values__tmp__table__3.FieldSchema(name:tmp_values_col4, type:string, comment:), ]
+POSTHOOK: Lineage: purchase_history.price SCRIPT []
+POSTHOOK: Lineage: purchase_history.product SCRIPT []
+POSTHOOK: Lineage: purchase_history.s SCRIPT []
+POSTHOOK: Lineage: purchase_history.time SCRIPT []
PREHOOK: query: insert into purchase_history values ('4', 'Shirt', 15.50, 59)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@purchase_history
POSTHOOK: query: insert into purchase_history values ('4', 'Shirt', 15.50, 59)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@purchase_history
-POSTHOOK: Lineage: purchase_history.price EXPRESSION [(values__tmp__table__4)values__tmp__table__4.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
-POSTHOOK: Lineage: purchase_history.product SIMPLE [(values__tmp__table__4)values__tmp__table__4.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: purchase_history.s SIMPLE [(values__tmp__table__4)values__tmp__table__4.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: purchase_history.time EXPRESSION [(values__tmp__table__4)values__tmp__table__4.FieldSchema(name:tmp_values_col4, type:string, comment:), ]
+POSTHOOK: Lineage: purchase_history.price SCRIPT []
+POSTHOOK: Lineage: purchase_history.product SCRIPT []
+POSTHOOK: Lineage: purchase_history.s SCRIPT []
+POSTHOOK: Lineage: purchase_history.time SCRIPT []
PREHOOK: query: create table cart_history (s string, cart_id int, `time` int)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
@@ -56,58 +64,70 @@ POSTHOOK: Output: database:default
POSTHOOK: Output: default@cart_history
PREHOOK: query: insert into cart_history values ('1', 1, 10)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@cart_history
POSTHOOK: query: insert into cart_history values ('1', 1, 10)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@cart_history
-POSTHOOK: Lineage: cart_history.cart_id EXPRESSION [(values__tmp__table__5)values__tmp__table__5.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: cart_history.s SIMPLE [(values__tmp__table__5)values__tmp__table__5.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: cart_history.time EXPRESSION [(values__tmp__table__5)values__tmp__table__5.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
+POSTHOOK: Lineage: cart_history.cart_id SCRIPT []
+POSTHOOK: Lineage: cart_history.s SCRIPT []
+POSTHOOK: Lineage: cart_history.time SCRIPT []
PREHOOK: query: insert into cart_history values ('1', 2, 20)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@cart_history
POSTHOOK: query: insert into cart_history values ('1', 2, 20)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@cart_history
-POSTHOOK: Lineage: cart_history.cart_id EXPRESSION [(values__tmp__table__6)values__tmp__table__6.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: cart_history.s SIMPLE [(values__tmp__table__6)values__tmp__table__6.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: cart_history.time EXPRESSION [(values__tmp__table__6)values__tmp__table__6.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
+POSTHOOK: Lineage: cart_history.cart_id SCRIPT []
+POSTHOOK: Lineage: cart_history.s SCRIPT []
+POSTHOOK: Lineage: cart_history.time SCRIPT []
PREHOOK: query: insert into cart_history values ('1', 3, 30)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@cart_history
POSTHOOK: query: insert into cart_history values ('1', 3, 30)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@cart_history
-POSTHOOK: Lineage: cart_history.cart_id EXPRESSION [(values__tmp__table__7)values__tmp__table__7.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: cart_history.s SIMPLE [(values__tmp__table__7)values__tmp__table__7.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: cart_history.time EXPRESSION [(values__tmp__table__7)values__tmp__table__7.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
+POSTHOOK: Lineage: cart_history.cart_id SCRIPT []
+POSTHOOK: Lineage: cart_history.s SCRIPT []
+POSTHOOK: Lineage: cart_history.time SCRIPT []
PREHOOK: query: insert into cart_history values ('1', 4, 40)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@cart_history
POSTHOOK: query: insert into cart_history values ('1', 4, 40)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@cart_history
-POSTHOOK: Lineage: cart_history.cart_id EXPRESSION [(values__tmp__table__8)values__tmp__table__8.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: cart_history.s SIMPLE [(values__tmp__table__8)values__tmp__table__8.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: cart_history.time EXPRESSION [(values__tmp__table__8)values__tmp__table__8.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
+POSTHOOK: Lineage: cart_history.cart_id SCRIPT []
+POSTHOOK: Lineage: cart_history.s SCRIPT []
+POSTHOOK: Lineage: cart_history.time SCRIPT []
PREHOOK: query: insert into cart_history values ('3', 5, 50)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@cart_history
POSTHOOK: query: insert into cart_history values ('3', 5, 50)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@cart_history
-POSTHOOK: Lineage: cart_history.cart_id EXPRESSION [(values__tmp__table__9)values__tmp__table__9.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: cart_history.s SIMPLE [(values__tmp__table__9)values__tmp__table__9.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: cart_history.time EXPRESSION [(values__tmp__table__9)values__tmp__table__9.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
+POSTHOOK: Lineage: cart_history.cart_id SCRIPT []
+POSTHOOK: Lineage: cart_history.s SCRIPT []
+POSTHOOK: Lineage: cart_history.time SCRIPT []
PREHOOK: query: insert into cart_history values ('4', 6, 60)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@cart_history
POSTHOOK: query: insert into cart_history values ('4', 6, 60)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@cart_history
-POSTHOOK: Lineage: cart_history.cart_id EXPRESSION [(values__tmp__table__10)values__tmp__table__10.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: cart_history.s SIMPLE [(values__tmp__table__10)values__tmp__table__10.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: cart_history.time EXPRESSION [(values__tmp__table__10)values__tmp__table__10.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
+POSTHOOK: Lineage: cart_history.cart_id SCRIPT []
+POSTHOOK: Lineage: cart_history.s SCRIPT []
+POSTHOOK: Lineage: cart_history.time SCRIPT []
PREHOOK: query: create table events (s string, st2 string, n int, `time` int)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
@@ -118,64 +138,76 @@ POSTHOOK: Output: database:default
POSTHOOK: Output: default@events
PREHOOK: query: insert into events values ('1', 'Bob', 1234, 20)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@events
POSTHOOK: query: insert into events values ('1', 'Bob', 1234, 20)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@events
-POSTHOOK: Lineage: events.n EXPRESSION [(values__tmp__table__11)values__tmp__table__11.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
-POSTHOOK: Lineage: events.s SIMPLE [(values__tmp__table__11)values__tmp__table__11.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: events.st2 SIMPLE [(values__tmp__table__11)values__tmp__table__11.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: events.time EXPRESSION [(values__tmp__table__11)values__tmp__table__11.FieldSchema(name:tmp_values_col4, type:string, comment:), ]
+POSTHOOK: Lineage: events.n SCRIPT []
+POSTHOOK: Lineage: events.s SCRIPT []
+POSTHOOK: Lineage: events.st2 SCRIPT []
+POSTHOOK: Lineage: events.time SCRIPT []
PREHOOK: query: insert into events values ('1', 'Bob', 1234, 30)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@events
POSTHOOK: query: insert into events values ('1', 'Bob', 1234, 30)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@events
-POSTHOOK: Lineage: events.n EXPRESSION [(values__tmp__table__12)values__tmp__table__12.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
-POSTHOOK: Lineage: events.s SIMPLE [(values__tmp__table__12)values__tmp__table__12.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: events.st2 SIMPLE [(values__tmp__table__12)values__tmp__table__12.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: events.time EXPRESSION [(values__tmp__table__12)values__tmp__table__12.FieldSchema(name:tmp_values_col4, type:string, comment:), ]
+POSTHOOK: Lineage: events.n SCRIPT []
+POSTHOOK: Lineage: events.s SCRIPT []
+POSTHOOK: Lineage: events.st2 SCRIPT []
+POSTHOOK: Lineage: events.time SCRIPT []
PREHOOK: query: insert into events values ('1', 'Bob', 1234, 25)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@events
POSTHOOK: query: insert into events values ('1', 'Bob', 1234, 25)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@events
-POSTHOOK: Lineage: events.n EXPRESSION [(values__tmp__table__13)values__tmp__table__13.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
-POSTHOOK: Lineage: events.s SIMPLE [(values__tmp__table__13)values__tmp__table__13.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: events.st2 SIMPLE [(values__tmp__table__13)values__tmp__table__13.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: events.time EXPRESSION [(values__tmp__table__13)values__tmp__table__13.FieldSchema(name:tmp_values_col4, type:string, comment:), ]
+POSTHOOK: Lineage: events.n SCRIPT []
+POSTHOOK: Lineage: events.s SCRIPT []
+POSTHOOK: Lineage: events.st2 SCRIPT []
+POSTHOOK: Lineage: events.time SCRIPT []
PREHOOK: query: insert into events values ('2', 'Sam', 1234, 30)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@events
POSTHOOK: query: insert into events values ('2', 'Sam', 1234, 30)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@events
-POSTHOOK: Lineage: events.n EXPRESSION [(values__tmp__table__14)values__tmp__table__14.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
-POSTHOOK: Lineage: events.s SIMPLE [(values__tmp__table__14)values__tmp__table__14.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: events.st2 SIMPLE [(values__tmp__table__14)values__tmp__table__14.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: events.time EXPRESSION [(values__tmp__table__14)values__tmp__table__14.FieldSchema(name:tmp_values_col4, type:string, comment:), ]
+POSTHOOK: Lineage: events.n SCRIPT []
+POSTHOOK: Lineage: events.s SCRIPT []
+POSTHOOK: Lineage: events.st2 SCRIPT []
+POSTHOOK: Lineage: events.time SCRIPT []
PREHOOK: query: insert into events values ('3', 'Jeff', 1234, 50)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@events
POSTHOOK: query: insert into events values ('3', 'Jeff', 1234, 50)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@events
-POSTHOOK: Lineage: events.n EXPRESSION [(values__tmp__table__15)values__tmp__table__15.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
-POSTHOOK: Lineage: events.s SIMPLE [(values__tmp__table__15)values__tmp__table__15.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: events.st2 SIMPLE [(values__tmp__table__15)values__tmp__table__15.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: events.time EXPRESSION [(values__tmp__table__15)values__tmp__table__15.FieldSchema(name:tmp_values_col4, type:string, comment:), ]
+POSTHOOK: Lineage: events.n SCRIPT []
+POSTHOOK: Lineage: events.s SCRIPT []
+POSTHOOK: Lineage: events.st2 SCRIPT []
+POSTHOOK: Lineage: events.time SCRIPT []
PREHOOK: query: insert into events values ('4', 'Ted', 1234, 60)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@events
POSTHOOK: query: insert into events values ('4', 'Ted', 1234, 60)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@events
-POSTHOOK: Lineage: events.n EXPRESSION [(values__tmp__table__16)values__tmp__table__16.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
-POSTHOOK: Lineage: events.s SIMPLE [(values__tmp__table__16)values__tmp__table__16.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: events.st2 SIMPLE [(values__tmp__table__16)values__tmp__table__16.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: events.time EXPRESSION [(values__tmp__table__16)values__tmp__table__16.FieldSchema(name:tmp_values_col4, type:string, comment:), ]
+POSTHOOK: Lineage: events.n SCRIPT []
+POSTHOOK: Lineage: events.s SCRIPT []
+POSTHOOK: Lineage: events.st2 SCRIPT []
+POSTHOOK: Lineage: events.time SCRIPT []
PREHOOK: query: explain
select s
from (
http://git-wip-us.apache.org/repos/asf/hive/blob/7e64114d/ql/src/test/results/clientpositive/join46.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/join46.q.out b/ql/src/test/results/clientpositive/join46.q.out
index e24d7b4..0847ca6 100644
--- a/ql/src/test/results/clientpositive/join46.q.out
+++ b/ql/src/test/results/clientpositive/join46.q.out
@@ -9,14 +9,16 @@ POSTHOOK: Output: default@test1
PREHOOK: query: INSERT INTO test1 VALUES (NULL, NULL, 'None'), (98, NULL, 'None'),
(99, 0, 'Alice'), (99, 2, 'Mat'), (100, 1, 'Bob'), (101, 2, 'Car')
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@test1
POSTHOOK: query: INSERT INTO test1 VALUES (NULL, NULL, 'None'), (98, NULL, 'None'),
(99, 0, 'Alice'), (99, 2, 'Mat'), (100, 1, 'Bob'), (101, 2, 'Car')
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@test1
-POSTHOOK: Lineage: test1.col_1 SIMPLE [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
-POSTHOOK: Lineage: test1.key EXPRESSION [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: test1.value EXPRESSION [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
+POSTHOOK: Lineage: test1.col_1 SCRIPT []
+POSTHOOK: Lineage: test1.key SCRIPT []
+POSTHOOK: Lineage: test1.value SCRIPT []
PREHOOK: query: CREATE TABLE test2 (key INT, value INT, col_2 STRING)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
@@ -28,14 +30,16 @@ POSTHOOK: Output: default@test2
PREHOOK: query: INSERT INTO test2 VALUES (102, 2, 'Del'), (103, 2, 'Ema'),
(104, 3, 'Fli'), (105, NULL, 'None')
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@test2
POSTHOOK: query: INSERT INTO test2 VALUES (102, 2, 'Del'), (103, 2, 'Ema'),
(104, 3, 'Fli'), (105, NULL, 'None')
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@test2
-POSTHOOK: Lineage: test2.col_2 SIMPLE [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
-POSTHOOK: Lineage: test2.key EXPRESSION [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: test2.value EXPRESSION [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
+POSTHOOK: Lineage: test2.col_2 SCRIPT []
+POSTHOOK: Lineage: test2.key SCRIPT []
+POSTHOOK: Lineage: test2.value SCRIPT []
PREHOOK: query: EXPLAIN
SELECT *
FROM test1 LEFT OUTER JOIN test2
http://git-wip-us.apache.org/repos/asf/hive/blob/7e64114d/ql/src/test/results/clientpositive/join_cond_pushdown_unqual5.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/join_cond_pushdown_unqual5.q.out b/ql/src/test/results/clientpositive/join_cond_pushdown_unqual5.q.out
index 203010f..5dd669f 100644
--- a/ql/src/test/results/clientpositive/join_cond_pushdown_unqual5.q.out
+++ b/ql/src/test/results/clientpositive/join_cond_pushdown_unqual5.q.out
@@ -16,23 +16,27 @@ POSTHOOK: Output: database:default
POSTHOOK: Output: default@rtable
PREHOOK: query: insert into ltable values (1, null, 'CD5415192314304', '00071'), (2, null, 'CD5415192225530', '00071')
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@ltable
POSTHOOK: query: insert into ltable values (1, null, 'CD5415192314304', '00071'), (2, null, 'CD5415192225530', '00071')
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@ltable
-POSTHOOK: Lineage: ltable.index EXPRESSION [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: ltable.la EXPRESSION [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: ltable.lk1 SIMPLE [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
-POSTHOOK: Lineage: ltable.lk2 SIMPLE [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col4, type:string, comment:), ]
+POSTHOOK: Lineage: ltable.index SCRIPT []
+POSTHOOK: Lineage: ltable.la EXPRESSION []
+POSTHOOK: Lineage: ltable.lk1 SCRIPT []
+POSTHOOK: Lineage: ltable.lk2 SCRIPT []
PREHOOK: query: insert into rtable values (1, 'CD5415192314304', '00071'), (45, 'CD5415192314304', '00072')
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@rtable
POSTHOOK: query: insert into rtable values (1, 'CD5415192314304', '00071'), (45, 'CD5415192314304', '00072')
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@rtable
-POSTHOOK: Lineage: rtable.ra EXPRESSION [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: rtable.rk1 SIMPLE [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: rtable.rk2 SIMPLE [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
+POSTHOOK: Lineage: rtable.ra SCRIPT []
+POSTHOOK: Lineage: rtable.rk1 SCRIPT []
+POSTHOOK: Lineage: rtable.rk2 SCRIPT []
PREHOOK: query: EXPLAIN SELECT * FROM ltable l LEFT OUTER JOIN rtable r on (l.lk1 = r.rk1 AND l.lk2 = r.rk2) WHERE COALESCE(l.la,'EMPTY')=COALESCE(r.ra,'EMPTY')
PREHOOK: type: QUERY
POSTHOOK: query: EXPLAIN SELECT * FROM ltable l LEFT OUTER JOIN rtable r on (l.lk1 = r.rk1 AND l.lk2 = r.rk2) WHERE COALESCE(l.la,'EMPTY')=COALESCE(r.ra,'EMPTY')
http://git-wip-us.apache.org/repos/asf/hive/blob/7e64114d/ql/src/test/results/clientpositive/join_emit_interval.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/join_emit_interval.q.out b/ql/src/test/results/clientpositive/join_emit_interval.q.out
index ce7c792..d28b15d 100644
--- a/ql/src/test/results/clientpositive/join_emit_interval.q.out
+++ b/ql/src/test/results/clientpositive/join_emit_interval.q.out
@@ -9,14 +9,16 @@ POSTHOOK: Output: default@test1
PREHOOK: query: INSERT INTO test1 VALUES (NULL, NULL, 'None'), (98, NULL, 'None'),
(99, 0, 'Alice'), (99, 2, 'Mat'), (100, 1, 'Bob'), (101, 2, 'Car')
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@test1
POSTHOOK: query: INSERT INTO test1 VALUES (NULL, NULL, 'None'), (98, NULL, 'None'),
(99, 0, 'Alice'), (99, 2, 'Mat'), (100, 1, 'Bob'), (101, 2, 'Car')
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@test1
-POSTHOOK: Lineage: test1.col_1 SIMPLE [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
-POSTHOOK: Lineage: test1.key EXPRESSION [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: test1.value EXPRESSION [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
+POSTHOOK: Lineage: test1.col_1 SCRIPT []
+POSTHOOK: Lineage: test1.key SCRIPT []
+POSTHOOK: Lineage: test1.value SCRIPT []
PREHOOK: query: CREATE TABLE test2 (key INT, value INT, col_2 STRING)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
@@ -28,14 +30,16 @@ POSTHOOK: Output: default@test2
PREHOOK: query: INSERT INTO test2 VALUES (102, 2, 'Del'), (103, 2, 'Ema'),
(104, 3, 'Fli'), (105, NULL, 'None')
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@test2
POSTHOOK: query: INSERT INTO test2 VALUES (102, 2, 'Del'), (103, 2, 'Ema'),
(104, 3, 'Fli'), (105, NULL, 'None')
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@test2
-POSTHOOK: Lineage: test2.col_2 SIMPLE [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
-POSTHOOK: Lineage: test2.key EXPRESSION [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: test2.value EXPRESSION [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
+POSTHOOK: Lineage: test2.col_2 SCRIPT []
+POSTHOOK: Lineage: test2.key SCRIPT []
+POSTHOOK: Lineage: test2.value SCRIPT []
PREHOOK: query: EXPLAIN
SELECT *
FROM test1 LEFT OUTER JOIN test2
http://git-wip-us.apache.org/repos/asf/hive/blob/7e64114d/ql/src/test/results/clientpositive/llap/acid_bucket_pruning.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/acid_bucket_pruning.q.out b/ql/src/test/results/clientpositive/llap/acid_bucket_pruning.q.out
index 77bea79..94486ec 100644
--- a/ql/src/test/results/clientpositive/llap/acid_bucket_pruning.q.out
+++ b/ql/src/test/results/clientpositive/llap/acid_bucket_pruning.q.out
@@ -17,11 +17,13 @@ POSTHOOK: Output: default@acidtbldefault
POSTHOOK: Lineage: acidtbldefault.a SIMPLE [(alltypesorc)alltypesorc.FieldSchema(name:cint, type:int, comment:null), ]
PREHOOK: query: INSERT INTO TABLE acidTblDefault VALUES (1)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@acidtbldefault
POSTHOOK: query: INSERT INTO TABLE acidTblDefault VALUES (1)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@acidtbldefault
-POSTHOOK: Lineage: acidtbldefault.a EXPRESSION [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
+POSTHOOK: Lineage: acidtbldefault.a SCRIPT []
PREHOOK: query: EXPLAIN EXTENDED
SELECT * FROM acidTblDefault WHERE a = 1
PREHOOK: type: QUERY
http://git-wip-us.apache.org/repos/asf/hive/blob/7e64114d/ql/src/test/results/clientpositive/llap/acid_no_buckets.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/acid_no_buckets.q.out b/ql/src/test/results/clientpositive/llap/acid_no_buckets.q.out
index 5fe7544..629b05d 100644
--- a/ql/src/test/results/clientpositive/llap/acid_no_buckets.q.out
+++ b/ql/src/test/results/clientpositive/llap/acid_no_buckets.q.out
@@ -210,12 +210,14 @@ POSTHOOK: Input: default@srcpart_acid@ds=2008-04-09/hr=12
2008-04-09 11 413 val_413updated
PREHOOK: query: insert into srcpart_acid PARTITION (ds='2008-04-08', hr=='11') values ('1001','val1001'),('1002','val1002'),('1003','val1003')
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@srcpart_acid@ds=2008-04-08/hr=11
POSTHOOK: query: insert into srcpart_acid PARTITION (ds='2008-04-08', hr=='11') values ('1001','val1001'),('1002','val1002'),('1003','val1003')
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@srcpart_acid@ds=2008-04-08/hr=11
-POSTHOOK: Lineage: srcpart_acid PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: srcpart_acid PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
+POSTHOOK: Lineage: srcpart_acid PARTITION(ds=2008-04-08,hr=11).key SCRIPT []
+POSTHOOK: Lineage: srcpart_acid PARTITION(ds=2008-04-08,hr=11).value SCRIPT []
PREHOOK: query: select ds, hr, key, value from srcpart_acid where cast(key as integer) > 1000 order by ds, hr, cast(key as integer)
PREHOOK: type: QUERY
PREHOOK: Input: default@srcpart_acid
@@ -708,12 +710,14 @@ POSTHOOK: Input: default@srcpart_acidb@ds=2008-04-09/hr=12
2008-04-09 11 413 val_413updated
PREHOOK: query: insert into srcpart_acidb PARTITION (ds='2008-04-08', hr=='11') values ('1001','val1001'),('1002','val1002'),('1003','val1003')
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@srcpart_acidb@ds=2008-04-08/hr=11
POSTHOOK: query: insert into srcpart_acidb PARTITION (ds='2008-04-08', hr=='11') values ('1001','val1001'),('1002','val1002'),('1003','val1003')
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@srcpart_acidb@ds=2008-04-08/hr=11
-POSTHOOK: Lineage: srcpart_acidb PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: srcpart_acidb PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
+POSTHOOK: Lineage: srcpart_acidb PARTITION(ds=2008-04-08,hr=11).key SCRIPT []
+POSTHOOK: Lineage: srcpart_acidb PARTITION(ds=2008-04-08,hr=11).value SCRIPT []
PREHOOK: query: select ds, hr, key, value from srcpart_acidb where cast(key as integer) > 1000 order by ds, hr, cast(key as integer)
PREHOOK: type: QUERY
PREHOOK: Input: default@srcpart_acidb
@@ -1223,12 +1227,14 @@ POSTHOOK: Input: default@srcpart_acidv@ds=2008-04-09/hr=12
2008-04-09 11 413 val_413updated
PREHOOK: query: insert into srcpart_acidv PARTITION (ds='2008-04-08', hr=='11') values ('1001','val1001'),('1002','val1002'),('1003','val1003')
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@srcpart_acidv@ds=2008-04-08/hr=11
POSTHOOK: query: insert into srcpart_acidv PARTITION (ds='2008-04-08', hr=='11') values ('1001','val1001'),('1002','val1002'),('1003','val1003')
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@srcpart_acidv@ds=2008-04-08/hr=11
-POSTHOOK: Lineage: srcpart_acidv PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(values__tmp__table__3)values__tmp__table__3.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: srcpart_acidv PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(values__tmp__table__3)values__tmp__table__3.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
+POSTHOOK: Lineage: srcpart_acidv PARTITION(ds=2008-04-08,hr=11).key SCRIPT []
+POSTHOOK: Lineage: srcpart_acidv PARTITION(ds=2008-04-08,hr=11).value SCRIPT []
PREHOOK: query: select ds, hr, key, value from srcpart_acidv where cast(key as integer) > 1000 order by ds, hr, cast(key as integer)
PREHOOK: type: QUERY
PREHOOK: Input: default@srcpart_acidv
@@ -2006,12 +2012,14 @@ POSTHOOK: Input: default@srcpart_acidvb@ds=2008-04-09/hr=12
2008-04-09 11 413 val_413updated
PREHOOK: query: insert into srcpart_acidvb PARTITION (ds='2008-04-08', hr=='11') values ('1001','val1001'),('1002','val1002'),('1003','val1003')
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@srcpart_acidvb@ds=2008-04-08/hr=11
POSTHOOK: query: insert into srcpart_acidvb PARTITION (ds='2008-04-08', hr=='11') values ('1001','val1001'),('1002','val1002'),('1003','val1003')
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@srcpart_acidvb@ds=2008-04-08/hr=11
-POSTHOOK: Lineage: srcpart_acidvb PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(values__tmp__table__4)values__tmp__table__4.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
-POSTHOOK: Lineage: srcpart_acidvb PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(values__tmp__table__4)values__tmp__table__4.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
+POSTHOOK: Lineage: srcpart_acidvb PARTITION(ds=2008-04-08,hr=11).key SCRIPT []
+POSTHOOK: Lineage: srcpart_acidvb PARTITION(ds=2008-04-08,hr=11).value SCRIPT []
PREHOOK: query: select ds, hr, key, value from srcpart_acidvb where cast(key as integer) > 1000 order by ds, hr, cast(key as integer)
PREHOOK: type: QUERY
PREHOOK: Input: default@srcpart_acidvb
http://git-wip-us.apache.org/repos/asf/hive/blob/7e64114d/ql/src/test/results/clientpositive/llap/acid_vectorization_original.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/acid_vectorization_original.q.out b/ql/src/test/results/clientpositive/llap/acid_vectorization_original.q.out
index 22a2375..68801f0 100644
--- a/ql/src/test/results/clientpositive/llap/acid_vectorization_original.q.out
+++ b/ql/src/test/results/clientpositive/llap/acid_vectorization_original.q.out
@@ -14,11 +14,13 @@ POSTHOOK: Output: database:default
POSTHOOK: Output: default@mydual
PREHOOK: query: insert into mydual values(1)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@mydual
POSTHOOK: query: insert into mydual values(1)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@mydual
-POSTHOOK: Lineage: mydual.a EXPRESSION [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
+POSTHOOK: Lineage: mydual.a SCRIPT []
PREHOOK: query: CREATE TABLE over10k(t tinyint,
si smallint,
i int,
http://git-wip-us.apache.org/repos/asf/hive/blob/7e64114d/ql/src/test/results/clientpositive/llap/authorization_view_8.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/authorization_view_8.q.out b/ql/src/test/results/clientpositive/llap/authorization_view_8.q.out
index 67ddb32..f4f5259 100644
--- a/ql/src/test/results/clientpositive/llap/authorization_view_8.q.out
+++ b/ql/src/test/results/clientpositive/llap/authorization_view_8.q.out
@@ -18,6 +18,7 @@ PREHOOK: query: insert into my_passwd values
("Gopal", 5),
("Sergey", 6)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@my_passwd
POSTHOOK: query: insert into my_passwd values
("Deepak", 1),
@@ -27,9 +28,10 @@ POSTHOOK: query: insert into my_passwd values
("Gopal", 5),
("Sergey", 6)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@my_passwd
-POSTHOOK: Lineage: my_passwd.uid EXPRESSION [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: my_passwd.username SIMPLE [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
+POSTHOOK: Lineage: my_passwd.uid SCRIPT []
+POSTHOOK: Lineage: my_passwd.username SCRIPT []
PREHOOK: query: create view my_passwd_vw as select * from my_passwd limit 3
PREHOOK: type: CREATEVIEW
PREHOOK: Input: default@my_passwd
http://git-wip-us.apache.org/repos/asf/hive/blob/7e64114d/ql/src/test/results/clientpositive/llap/autoColumnStats_10.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/autoColumnStats_10.q.out b/ql/src/test/results/clientpositive/llap/autoColumnStats_10.q.out
index 6cb51fd..9ae59e6 100644
--- a/ql/src/test/results/clientpositive/llap/autoColumnStats_10.q.out
+++ b/ql/src/test/results/clientpositive/llap/autoColumnStats_10.q.out
@@ -47,13 +47,15 @@ Storage Desc Params:
serialization.format 1
PREHOOK: query: insert into p values (1,22,333)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@p
POSTHOOK: query: insert into p values (1,22,333)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@p
-POSTHOOK: Lineage: p.c1 EXPRESSION [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: p.c2 EXPRESSION [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
-POSTHOOK: Lineage: p.insert_num EXPRESSION [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
+POSTHOOK: Lineage: p.c1 SCRIPT []
+POSTHOOK: Lineage: p.c2 SCRIPT []
+POSTHOOK: Lineage: p.insert_num SCRIPT []
PREHOOK: query: desc formatted p
PREHOOK: type: DESCTABLE
PREHOOK: Input: default@p
@@ -173,13 +175,15 @@ comment from deserializer
COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"insert_num\":\"true\"}}
PREHOOK: query: insert into p values (2,11,111)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@p
POSTHOOK: query: insert into p values (2,11,111)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@p
-POSTHOOK: Lineage: p.c1 SIMPLE [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: p.c2 SIMPLE [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
-POSTHOOK: Lineage: p.insert_num EXPRESSION [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
+POSTHOOK: Lineage: p.c1 SCRIPT []
+POSTHOOK: Lineage: p.c2 SCRIPT []
+POSTHOOK: Lineage: p.insert_num SCRIPT []
PREHOOK: query: desc formatted p
PREHOOK: type: DESCTABLE
PREHOOK: Input: default@p
@@ -307,13 +311,15 @@ Storage Desc Params:
serialization.format 1
PREHOOK: query: insert into p values (1,22,333)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@p
POSTHOOK: query: insert into p values (1,22,333)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@p
-POSTHOOK: Lineage: p.c1 EXPRESSION [(values__tmp__table__3)values__tmp__table__3.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: p.c2 EXPRESSION [(values__tmp__table__3)values__tmp__table__3.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
-POSTHOOK: Lineage: p.insert_num EXPRESSION [(values__tmp__table__3)values__tmp__table__3.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
+POSTHOOK: Lineage: p.c1 SCRIPT []
+POSTHOOK: Lineage: p.c2 SCRIPT []
+POSTHOOK: Lineage: p.insert_num SCRIPT []
PREHOOK: query: desc formatted p
PREHOOK: type: DESCTABLE
PREHOOK: Input: default@p
@@ -433,13 +439,15 @@ comment from deserializer
COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
PREHOOK: query: insert into p values (2,11,111)
PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@p
POSTHOOK: query: insert into p values (2,11,111)
POSTHOOK: type: QUERY
+POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@p
-POSTHOOK: Lineage: p.c1 SIMPLE [(values__tmp__table__4)values__tmp__table__4.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
-POSTHOOK: Lineage: p.c2 SIMPLE [(values__tmp__table__4)values__tmp__table__4.FieldSchema(name:tmp_values_col3, type:string, comment:), ]
-POSTHOOK: Lineage: p.insert_num EXPRESSION [(values__tmp__table__4)values__tmp__table__4.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
+POSTHOOK: Lineage: p.c1 SCRIPT []
+POSTHOOK: Lineage: p.c2 SCRIPT []
+POSTHOOK: Lineage: p.insert_num SCRIPT []
PREHOOK: query: desc formatted p
PREHOOK: type: DESCTABLE
PREHOOK: Input: default@p
http://git-wip-us.apache.org/repos/asf/hive/blob/7e64114d/ql/src/test/results/clientpositive/llap/bucket_map_join_tez1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/bucket_map_join_tez1.q.out b/ql/src/test/results/clientpositive/llap/bucket_map_join_tez1.q.out
index 543fccd..19a503e 100644
--- a/ql/src/test/results/clientpositive/llap/bucket_map_join_tez1.q.out
+++ b/ql/src/test/results/clientpositive/llap/bucket_map_join_tez1.q.out
@@ -235,25 +235,25 @@ STAGE PLANS:
0 _col0 (type: int)
1 _col0 (type: int)
outputColumnNames: _col0, _col1, _col3
- Statistics: Num rows: 391 Data size: 72726 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 72912 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col0 (type: int), _col1 (type: string), _col3 (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 391 Data size: 72726 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 72912 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
key expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string)
sort order: +++
- Statistics: Num rows: 391 Data size: 72726 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 72912 Basic stats: COMPLETE Column stats: COMPLETE
Reducer 3
Execution mode: llap
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 391 Data size: 72726 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 72912 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 391 Data size: 72726 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 72912 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -823,15 +823,15 @@ STAGE PLANS:
outputColumnNames: _col0, _col1, _col3
input vertices:
0 Map 1
- Statistics: Num rows: 391 Data size: 72726 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 72912 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col0 (type: int), _col1 (type: string), _col3 (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 391 Data size: 72726 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 72912 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
key expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string)
sort order: +++
- Statistics: Num rows: 391 Data size: 72726 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 72912 Basic stats: COMPLETE Column stats: COMPLETE
Execution mode: llap
LLAP IO: no inputs
Reducer 3
@@ -840,10 +840,10 @@ STAGE PLANS:
Select Operator
expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 391 Data size: 72726 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 72912 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 391 Data size: 72726 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 72912 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -1436,7 +1436,7 @@ STAGE PLANS:
keys:
0 _col0 (type: int)
1 _col0 (type: int)
- Statistics: Num rows: 195 Data size: 1560 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 196 Data size: 1568 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: count()
mode: hash
@@ -1565,7 +1565,7 @@ STAGE PLANS:
1 _col0 (type: int)
input vertices:
1 Map 4
- Statistics: Num rows: 195 Data size: 1560 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 196 Data size: 1568 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: count()
mode: hash
@@ -1696,7 +1696,7 @@ STAGE PLANS:
keys:
0 _col0 (type: int)
1 _col0 (type: int)
- Statistics: Num rows: 632 Data size: 5056 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 636 Data size: 5088 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: count()
mode: hash
@@ -1731,16 +1731,16 @@ STAGE PLANS:
0 _col0 (type: int)
1 _col0 (type: int)
outputColumnNames: _col1
- Statistics: Num rows: 391 Data size: 1564 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 1568 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col1 (type: int)
outputColumnNames: _col0
- Statistics: Num rows: 391 Data size: 1564 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 1568 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
key expressions: _col0 (type: int)
sort order: +
Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 391 Data size: 1564 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 1568 Basic stats: COMPLETE Column stats: COMPLETE
Stage: Stage-0
Fetch Operator
@@ -1838,11 +1838,11 @@ STAGE PLANS:
outputColumnNames: _col1
input vertices:
1 Map 4
- Statistics: Num rows: 391 Data size: 1564 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 1568 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col1 (type: int)
outputColumnNames: _col0
- Statistics: Num rows: 391 Data size: 1564 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 1568 Basic stats: COMPLETE Column stats: COMPLETE
Map Join Operator
condition map:
Inner Join 0 to 1
@@ -1851,7 +1851,7 @@ STAGE PLANS:
1 _col0 (type: int)
input vertices:
0 Map 1
- Statistics: Num rows: 632 Data size: 5056 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 636 Data size: 5088 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: count()
mode: hash
@@ -2007,7 +2007,7 @@ STAGE PLANS:
keys:
0 _col0 (type: int)
1 _col0 (type: int)
- Statistics: Num rows: 632 Data size: 5056 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 636 Data size: 5088 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: count()
mode: hash
@@ -2042,16 +2042,16 @@ STAGE PLANS:
0 _col0 (type: int)
1 _col0 (type: int)
outputColumnNames: _col1
- Statistics: Num rows: 391 Data size: 1564 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 1568 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col1 (type: int)
outputColumnNames: _col0
- Statistics: Num rows: 391 Data size: 1564 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 1568 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
key expressions: _col0 (type: int)
sort order: +
Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 391 Data size: 1564 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 1568 Basic stats: COMPLETE Column stats: COMPLETE
Stage: Stage-0
Fetch Operator
@@ -2149,11 +2149,11 @@ STAGE PLANS:
outputColumnNames: _col1
input vertices:
1 Map 4
- Statistics: Num rows: 391 Data size: 1564 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 1568 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col1 (type: int)
outputColumnNames: _col0
- Statistics: Num rows: 391 Data size: 1564 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 1568 Basic stats: COMPLETE Column stats: COMPLETE
Map Join Operator
condition map:
Inner Join 0 to 1
@@ -2162,7 +2162,7 @@ STAGE PLANS:
1 _col0 (type: int)
input vertices:
0 Map 1
- Statistics: Num rows: 632 Data size: 5056 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 636 Data size: 5088 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: count()
mode: hash
@@ -2566,11 +2566,11 @@ STAGE PLANS:
0 _col0 (type: int)
1 _col0 (type: int)
outputColumnNames: _col1, _col2
- Statistics: Num rows: 391 Data size: 37145 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 37240 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col1 (type: int), substr(_col2, 5) (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 391 Data size: 37145 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 37240 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: sum(_col1)
keys: _col0 (type: int)
@@ -2708,11 +2708,11 @@ STAGE PLANS:
outputColumnNames: _col1, _col2
input vertices:
0 Map 2
- Statistics: Num rows: 391 Data size: 37145 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 37240 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col1 (type: int), substr(_col2, 5) (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 391 Data size: 37145 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 37240 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: sum(_col1)
keys: _col0 (type: int)
@@ -2866,14 +2866,14 @@ STAGE PLANS:
0 _col0 (type: int)
1 _col1 (type: int)
outputColumnNames: _col1, _col2, _col3
- Statistics: Num rows: 247 Data size: 25441 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 248 Data size: 25544 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col3 (type: int), _col2 (type: double), _col1 (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 247 Data size: 25441 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 248 Data size: 25544 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 247 Data size: 25441 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 248 Data size: 25544 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -2949,14 +2949,14 @@ STAGE PLANS:
outputColumnNames: _col1, _col2, _col3
input vertices:
1 Reducer 3
- Statistics: Num rows: 247 Data size: 25441 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 248 Data size: 25544 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col3 (type: int), _col2 (type: double), _col1 (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 247 Data size: 25441 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 248 Data size: 25544 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 247 Data size: 25441 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 248 Data size: 25544 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -3139,14 +3139,14 @@ STAGE PLANS:
1 _col0 (type: int)
2 _col0 (type: int)
outputColumnNames: _col0, _col1, _col4
- Statistics: Num rows: 619 Data size: 115134 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 621 Data size: 115506 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col0 (type: int), _col1 (type: string), _col4 (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 619 Data size: 115134 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 621 Data size: 115506 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 619 Data size: 115134 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 621 Data size: 115506 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -3202,14 +3202,14 @@ STAGE PLANS:
input vertices:
1 Map 2
2 Map 3
- Statistics: Num rows: 619 Data size: 115134 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 621 Data size: 115506 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col0 (type: int), _col1 (type: string), _col4 (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 619 Data size: 115134 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 621 Data size: 115506 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 619 Data size: 115134 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 621 Data size: 115506 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -3304,12 +3304,12 @@ STAGE PLANS:
outputColumnNames: _col0, _col1
input vertices:
1 Map 3
- Statistics: Num rows: 391 Data size: 37145 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 37240 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
key expressions: _col0 (type: int)
sort order: +
Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 391 Data size: 37145 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 37240 Basic stats: COMPLETE Column stats: COMPLETE
value expressions: _col1 (type: string)
Execution mode: llap
LLAP IO: no inputs
@@ -3362,14 +3362,14 @@ STAGE PLANS:
0 _col0 (type: int)
1 _col0 (type: int)
outputColumnNames: _col0, _col1, _col3
- Statistics: Num rows: 618 Data size: 114948 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 620 Data size: 115320 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col0 (type: int), _col1 (type: string), _col3 (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 618 Data size: 114948 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 620 Data size: 115320 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 618 Data size: 114948 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 620 Data size: 115320 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -3422,7 +3422,7 @@ STAGE PLANS:
outputColumnNames: _col0, _col1
input vertices:
1 Map 2
- Statistics: Num rows: 391 Data size: 37145 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 37240 Basic stats: COMPLETE Column stats: COMPLETE
Map Join Operator
condition map:
Inner Join 0 to 1
@@ -3432,14 +3432,14 @@ STAGE PLANS:
outputColumnNames: _col0, _col1, _col3
input vertices:
1 Map 3
- Statistics: Num rows: 618 Data size: 114948 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 620 Data size: 115320 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col0 (type: int), _col1 (type: string), _col3 (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 618 Data size: 114948 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 620 Data size: 115320 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 618 Data size: 114948 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 620 Data size: 115320 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -3586,14 +3586,14 @@ STAGE PLANS:
0 _col0 (type: int)
1 _col0 (type: int)
outputColumnNames: _col0, _col1, _col3
- Statistics: Num rows: 195 Data size: 20085 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 196 Data size: 20188 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col0 (type: int), _col1 (type: double), _col3 (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 195 Data size: 20085 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 196 Data size: 20188 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 195 Data size: 20085 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 196 Data size: 20188 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -3675,14 +3675,14 @@ STAGE PLANS:
outputColumnNames: _col0, _col1, _col3
input vertices:
0 Reducer 2
- Statistics: Num rows: 195 Data size: 20085 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 196 Data size: 20188 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col0 (type: int), _col1 (type: double), _col3 (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 195 Data size: 20085 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 196 Data size: 20188 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 195 Data size: 20085 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 196 Data size: 20188 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -3799,14 +3799,14 @@ STAGE PLANS:
0 _col0 (type: int)
1 _col0 (type: int)
outputColumnNames: _col0, _col1, _col3
- Statistics: Num rows: 247 Data size: 25441 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 248 Data size: 25544 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col0 (type: int), _col1 (type: double), _col3 (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 247 Data size: 25441 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 248 Data size: 25544 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 247 Data size: 25441 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 248 Data size: 25544 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -3882,14 +3882,14 @@ STAGE PLANS:
outputColumnNames: _col0, _col1, _col3
input vertices:
0 Reducer 2
- Statistics: Num rows: 247 Data size: 25441 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 248 Data size: 25544 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col0 (type: int), _col1 (type: double), _col3 (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 247 Data size: 25441 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 248 Data size: 25544 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 247 Data size: 25441 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 248 Data size: 25544 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -4178,14 +4178,14 @@ STAGE PLANS:
0 _col0 (type: int)
1 _col0 (type: int)
outputColumnNames: _col0, _col1, _col3
- Statistics: Num rows: 391 Data size: 72726 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 72912 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col0 (type: int), _col1 (type: string), _col3 (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 391 Data size: 72726 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 72912 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 391 Data size: 72726 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 72912 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -4258,14 +4258,14 @@ STAGE PLANS:
outputColumnNames: _col0, _col1, _col3
input vertices:
0 Map 1
- Statistics: Num rows: 391 Data size: 72726 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 72912 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col0 (type: int), _col1 (type: string), _col3 (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 391 Data size: 72726 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 72912 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 391 Data size: 72726 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 72912 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -4293,7 +4293,7 @@ STAGE PLANS:
#### A masked pattern was here ####
Edges:
Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE)
- Reducer 3 <- Map 1 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE)
+ Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE)
#### A masked pattern was here ####
Vertices:
Map 1
@@ -4309,11 +4309,18 @@ STAGE PLANS:
outputColumnNames: _col0, _col1
Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
- key expressions: _col0 (type: int)
+ key expressions: _col1 (type: string)
sort order: +
- Map-reduce partition columns: _col0 (type: int)
+ Map-reduce partition columns: _col1 (type: string)
Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
- value expressions: _col1 (type: string)
+ value expressions: _col0 (type: int)
+ Execution mode: llap
+ LLAP IO: no inputs
+ Map 4
+ Map Operator Tree:
+ TableScan
+ alias: b
+ Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
predicate: value is not null (type: boolean)
Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
@@ -4329,7 +4336,7 @@ STAGE PLANS:
value expressions: _col0 (type: int)
Execution mode: llap
LLAP IO: no inputs
- Map 4
+ Map 5
Map Operator Tree:
TableScan
alias: c
@@ -4355,16 +4362,16 @@ STAGE PLANS:
condition map:
Inner Join 0 to 1
keys:
- 0 _col0 (type: int)
- 1 _col0 (type: int)
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 809 Data size: 76855 Basic stats: COMPLETE Column stats: COMPLETE
+ 0 _col1 (type: string)
+ 1 _col1 (type: string)
+ outputColumnNames: _col0, _col2
+ Statistics: Num rows: 809 Data size: 6472 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
- key expressions: _col1 (type: string)
+ key expressions: _col0 (type: int)
sort order: +
- Map-reduce partition columns: _col1 (type: string)
- Statistics: Num rows: 809 Data size: 76855 Basic stats: COMPLETE Column stats: COMPLETE
- value expressions: _col0 (type: int)
+ Map-reduce partition columns: _col0 (type: int)
+ Statistics: Num rows: 809 Data size: 6472 Basic stats: COMPLETE Column stats: COMPLETE
+ value expressions: _col2 (type: int)
Reducer 3
Execution mode: llap
Reduce Operator Tree:
@@ -4372,17 +4379,17 @@ STAGE PLANS:
condition map:
Inner Join 0 to 1
keys:
- 0 _col1 (type: string)
- 1 _col1 (type: string)
- outputColumnNames: _col0, _col3
- Statistics: Num rows: 1309 Data size: 10472 Basic stats: COMPLETE Column stats: COMPLETE
+ 0 _col0 (type: int)
+ 1 _col0 (type: int)
+ outputColumnNames: _col0, _col2
+ Statistics: Num rows: 1313 Data size: 10504 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
- expressions: _col0 (type: int), _col3 (type: int)
+ expressions: _col0 (type: int), _col2 (type: int)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1309 Data size: 10472 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 1313 Data size: 10504 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 1309 Data size: 10472 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 1313 Data size: 10504 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -4407,8 +4414,8 @@ STAGE PLANS:
Tez
#### A masked pattern was here ####
Edges:
- Map 1 <- Map 3 (CUSTOM_EDGE)
Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE)
+ Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE)
#### A masked pattern was here ####
Vertices:
Map 1
@@ -4423,41 +4430,12 @@ STAGE PLANS:
expressions: key (type: int), value (type: string)
outputColumnNames: _col0, _col1
Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
- Map Join Operator
- condition map:
- Inner Join 0 to 1
- keys:
- 0 _col0 (type: int)
- 1 _col0 (type: int)
- outputColumnNames: _col0, _col1
- input vertices:
- 1 Map 3
- Statistics: Num rows: 809 Data size: 76855 Basic stats: COMPLETE Column stats: COMPLETE
- Reduce Output Operator
- key expressions: _col1 (type: string)
- sort order: +
- Map-reduce partition columns: _col1 (type: string)
- Statistics: Num rows: 809 Data size: 76855 Basic stats: COMPLETE Column stats: COMPLETE
- value expressions: _col0 (type: int)
- Execution mode: llap
- LLAP IO: no inputs
- Map 3
- Map Operator Tree:
- TableScan
- alias: c
- Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE
- Filter Operator
- predicate: key is not null (type: boolean)
- Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE
- Select Operator
- expressions: key (type: int)
- outputColumnNames: _col0
- Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
- key expressions: _col0 (type: int)
+ key expressions: _col1 (type: string)
sort order: +
- Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE
+ Map-reduce partition columns: _col1 (type: string)
+ Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
+ value expressions: _col0 (type: int)
Execution mode: llap
LLAP IO: no inputs
Map 4
@@ -4480,6 +4458,25 @@ STAGE PLANS:
value expressions: _col0 (type: int)
Execution mode: llap
LLAP IO: no inputs
+ Map 5
+ Map Operator Tree:
+ TableScan
+ alias: c
+ Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE
+ Filter Operator
+ predicate: key is not null (type: boolean)
+ Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE
+ Select Operator
+ expressions: key (type: int)
+ outputColumnNames: _col0
+ Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE
+ Reduce Output Operator
+ key expressions: _col0 (type: int)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: int)
+ Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE
+ Execution mode: llap
+ LLAP IO: no inputs
Reducer 2
Execution mode: llap
Reduce Operator Tree:
@@ -4489,15 +4486,32 @@ STAGE PLANS:
keys:
0 _col1 (type: string)
1 _col1 (type: string)
- outputColumnNames: _col0, _col3
- Statistics: Num rows: 1309 Data size: 10472 Basic stats: COMPLETE Column stats: COMPLETE
+ outputColumnNames: _col0, _col2
+ Statistics: Num rows: 809 Data size: 6472 Basic stats: COMPLETE Column stats: COMPLETE
+ Reduce Output Operator
+ key expressions: _col0 (type: int)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: int)
+ Statistics: Num rows: 809 Data size: 6472 Basic stats: COMPLETE Column stats: COMPLETE
+ value expressions: _col2 (type: int)
+ Reducer 3
+ Execution mode: llap
+ Reduce Operator Tree:
+ Merge Join Operator
+ condition map:
+ Inner Join 0 to 1
+ keys:
+ 0 _col0 (type: int)
+ 1 _col0 (type: int)
+ outputColumnNames: _col0, _col2
+ Statistics: Num rows: 1313 Data size: 10504 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
- expressions: _col0 (type: int), _col3 (type: int)
+ expressions: _col0 (type: int), _col2 (type: int)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1309 Data size: 10472 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 1313 Data size: 10504 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 1309 Data size: 10472 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 1313 Data size: 10504 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -4594,14 +4608,14 @@ STAGE PLANS:
0 _col0 (type: int), _col2 (type: string)
1 _col0 (type: int), _col2 (type: string)
outputColumnNames: _col0, _col1, _col4
- Statistics: Num rows: 391 Data size: 72726 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 72912 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col0 (type: int), _col1 (type: string), _col4 (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 391 Data size: 72726 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 72912 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 391 Data size: 72726 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 72912 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -4698,14 +4712,14 @@ STAGE PLANS:
0 _col0 (type: int), _col2 (type: string)
1 _col0 (type: int), _col2 (type: string)
outputColumnNames: _col0, _col1, _col4
- Statistics: Num rows: 391 Data size: 72726 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 72912 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col0 (type: int), _col1 (type: string), _col4 (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 391 Data size: 72726 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 72912 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 391 Data size: 72726 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 392 Data size: 72912 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat