You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by kg...@apache.org on 2018/02/27 07:01:36 UTC
[6/9] hive git commit: HIVE-18759: Remove unconnected q.out-s (Zoltan
Haindrich reviewed by Ashutosh Chauhan)
http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/llap/vector_gather_stats.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_gather_stats.q.out b/ql/src/test/results/clientpositive/llap/vector_gather_stats.q.out
deleted file mode 100644
index e777242..0000000
--- a/ql/src/test/results/clientpositive/llap/vector_gather_stats.q.out
+++ /dev/null
@@ -1,108 +0,0 @@
-PREHOOK: query: create table cd
-(
- cd_demo_sk int,
- cd_gender string,
- cd_marital_status string,
- cd_purchase_estimate int,
- cd_credit_rating string,
- cd_dep_count int,
- cd_dep_employed_count int,
- cd_dep_college_count int
-)
-partitioned by
-(
- cd_education_status string
-)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@cd
-POSTHOOK: query: create table cd
-(
- cd_demo_sk int,
- cd_gender string,
- cd_marital_status string,
- cd_purchase_estimate int,
- cd_credit_rating string,
- cd_dep_count int,
- cd_dep_employed_count int,
- cd_dep_college_count int
-)
-partitioned by
-(
- cd_education_status string
-)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@cd
-PREHOOK: query: alter table cd add partition (cd_education_status='Primary')
-PREHOOK: type: ALTERTABLE_ADDPARTS
-PREHOOK: Output: default@cd
-POSTHOOK: query: alter table cd add partition (cd_education_status='Primary')
-POSTHOOK: type: ALTERTABLE_ADDPARTS
-POSTHOOK: Output: default@cd
-POSTHOOK: Output: default@cd@cd_education_status=Primary
-PREHOOK: query: insert into table cd partition (cd_education_status='Primary') values (1, 'M', 'M', 500, 'Good', 0, 0, 0)
-PREHOOK: type: QUERY
-PREHOOK: Input: _dummy_database@_dummy_table
-PREHOOK: Output: default@cd@cd_education_status=Primary
-POSTHOOK: query: insert into table cd partition (cd_education_status='Primary') values (1, 'M', 'M', 500, 'Good', 0, 0, 0)
-POSTHOOK: type: QUERY
-POSTHOOK: Input: _dummy_database@_dummy_table
-POSTHOOK: Output: default@cd@cd_education_status=Primary
-POSTHOOK: Lineage: cd PARTITION(cd_education_status=Primary).cd_credit_rating SCRIPT []
-POSTHOOK: Lineage: cd PARTITION(cd_education_status=Primary).cd_demo_sk SCRIPT []
-POSTHOOK: Lineage: cd PARTITION(cd_education_status=Primary).cd_dep_college_count SCRIPT []
-POSTHOOK: Lineage: cd PARTITION(cd_education_status=Primary).cd_dep_count SCRIPT []
-POSTHOOK: Lineage: cd PARTITION(cd_education_status=Primary).cd_dep_employed_count SCRIPT []
-POSTHOOK: Lineage: cd PARTITION(cd_education_status=Primary).cd_gender SCRIPT []
-POSTHOOK: Lineage: cd PARTITION(cd_education_status=Primary).cd_marital_status SCRIPT []
-POSTHOOK: Lineage: cd PARTITION(cd_education_status=Primary).cd_purchase_estimate SCRIPT []
-PREHOOK: query: explain vectorization detail
-analyze table cd partition (cd_education_status) compute statistics
-PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization detail
-analyze table cd partition (cd_education_status) compute statistics
-POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
- enabled: true
- enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
-STAGE DEPENDENCIES:
- Stage-0 is a root stage
- Stage-2 depends on stages: Stage-0
-
-STAGE PLANS:
- Stage: Stage-0
- Tez
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: cd
- Statistics: Num rows: 1 Data size: 28 Basic stats: COMPLETE Column stats: NONE
- Execution mode: llap
- LLAP IO: no inputs
- Map Vectorization:
- enabled: true
- enabledConditionsMet: hive.vectorized.use.vector.serde.deserialize IS true
- inputFileFormats: org.apache.hadoop.mapred.TextInputFormat
- notVectorizedReason: TABLESCAN operator: gather stats not supported
- vectorized: false
-
- Stage: Stage-2
- Stats Work
- Basic Stats Work:
-
-PREHOOK: query: analyze table cd partition (cd_education_status) compute statistics
-PREHOOK: type: QUERY
-PREHOOK: Input: default@cd
-PREHOOK: Input: default@cd@cd_education_status=Primary
-PREHOOK: Output: default@cd
-PREHOOK: Output: default@cd@cd_education_status=Primary
-POSTHOOK: query: analyze table cd partition (cd_education_status) compute statistics
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@cd
-POSTHOOK: Input: default@cd@cd_education_status=Primary
-POSTHOOK: Output: default@cd
-POSTHOOK: Output: default@cd@cd_education_status=Primary
http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/llap/vector_join.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_join.q.out b/ql/src/test/results/clientpositive/llap/vector_join.q.out
deleted file mode 100644
index 94c0290..0000000
--- a/ql/src/test/results/clientpositive/llap/vector_join.q.out
+++ /dev/null
@@ -1,104 +0,0 @@
-PREHOOK: query: DROP TABLE IF EXISTS test1_vc
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE IF EXISTS test1_vc
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: DROP TABLE IF EXISTS test2_vc
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE IF EXISTS test2_vc
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE test1_vc
- (
- id string)
- PARTITIONED BY (
- cr_year bigint,
- cr_month bigint)
- ROW FORMAT SERDE
- 'org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe'
-STORED AS INPUTFORMAT
- 'org.apache.hadoop.hive.ql.io.RCFileInputFormat'
-OUTPUTFORMAT
- 'org.apache.hadoop.hive.ql.io.RCFileOutputFormat'
-TBLPROPERTIES (
- 'serialization.null.format'='' )
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@test1_vc
-POSTHOOK: query: CREATE TABLE test1_vc
- (
- id string)
- PARTITIONED BY (
- cr_year bigint,
- cr_month bigint)
- ROW FORMAT SERDE
- 'org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe'
-STORED AS INPUTFORMAT
- 'org.apache.hadoop.hive.ql.io.RCFileInputFormat'
-OUTPUTFORMAT
- 'org.apache.hadoop.hive.ql.io.RCFileOutputFormat'
-TBLPROPERTIES (
- 'serialization.null.format'='' )
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@test1_vc
-PREHOOK: query: CREATE TABLE test2_vc(
- id string
- )
- PARTITIONED BY (
- cr_year bigint,
- cr_month bigint)
-ROW FORMAT SERDE
- 'org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe'
-STORED AS INPUTFORMAT
- 'org.apache.hadoop.hive.ql.io.RCFileInputFormat'
-OUTPUTFORMAT
- 'org.apache.hadoop.hive.ql.io.RCFileOutputFormat'
-TBLPROPERTIES (
- 'serialization.null.format'=''
- )
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@test2_vc
-POSTHOOK: query: CREATE TABLE test2_vc(
- id string
- )
- PARTITIONED BY (
- cr_year bigint,
- cr_month bigint)
-ROW FORMAT SERDE
- 'org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe'
-STORED AS INPUTFORMAT
- 'org.apache.hadoop.hive.ql.io.RCFileInputFormat'
-OUTPUTFORMAT
- 'org.apache.hadoop.hive.ql.io.RCFileOutputFormat'
-TBLPROPERTIES (
- 'serialization.null.format'=''
- )
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@test2_vc
-PREHOOK: query: SELECT cr.id1 ,
-cr.id2
-FROM
-(SELECT t1.id id1,
- t2.id id2
- from
- (select * from test1_vc ) t1
- left outer join test2_vc t2
- on t1.id=t2.id) cr
-PREHOOK: type: QUERY
-PREHOOK: Input: default@test1_vc
-PREHOOK: Input: default@test2_vc
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT cr.id1 ,
-cr.id2
-FROM
-(SELECT t1.id id1,
- t2.id id2
- from
- (select * from test1_vc ) t1
- left outer join test2_vc t2
- on t1.id=t2.id) cr
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@test1_vc
-POSTHOOK: Input: default@test2_vc
-#### A masked pattern was here ####
http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/llap/vector_join_part_col_char.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_join_part_col_char.q.out b/ql/src/test/results/clientpositive/llap/vector_join_part_col_char.q.out
deleted file mode 100644
index 5e168a9..0000000
--- a/ql/src/test/results/clientpositive/llap/vector_join_part_col_char.q.out
+++ /dev/null
@@ -1,175 +0,0 @@
-PREHOOK: query: drop table if exists char_part_tbl1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table if exists char_part_tbl1
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: drop table if exists char_part_tbl2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table if exists char_part_tbl2
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: create table studenttab(name string, age int, gpa double) clustered by (age) into 2 buckets stored as orc tblproperties('transactional'='true')
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@studenttab
-POSTHOOK: query: create table studenttab(name string, age int, gpa double) clustered by (age) into 2 buckets stored as orc tblproperties('transactional'='true')
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@studenttab
-PREHOOK: query: insert into table studenttab values ('calvin garcia',56,2.50), ('oscar miller',66,3.00), ('(yuri xylophone',30,2.74),('alice underhill',46,3.50)
-PREHOOK: type: QUERY
-PREHOOK: Input: _dummy_database@_dummy_table
-PREHOOK: Output: default@studenttab
-POSTHOOK: query: insert into table studenttab values ('calvin garcia',56,2.50), ('oscar miller',66,3.00), ('(yuri xylophone',30,2.74),('alice underhill',46,3.50)
-POSTHOOK: type: QUERY
-POSTHOOK: Input: _dummy_database@_dummy_table
-POSTHOOK: Output: default@studenttab
-POSTHOOK: Lineage: studenttab.age SCRIPT []
-POSTHOOK: Lineage: studenttab.gpa SCRIPT []
-POSTHOOK: Lineage: studenttab.name SCRIPT []
-PREHOOK: query: create table char_tbl1(name string, age int) partitioned by(gpa char(50)) stored as orc
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@char_tbl1
-POSTHOOK: query: create table char_tbl1(name string, age int) partitioned by(gpa char(50)) stored as orc
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@char_tbl1
-PREHOOK: query: create table char_tbl2(name string, age int) partitioned by(gpa char(5)) stored as orc
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@char_tbl2
-POSTHOOK: query: create table char_tbl2(name string, age int) partitioned by(gpa char(5)) stored as orc
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@char_tbl2
-PREHOOK: query: insert into table char_tbl1 partition(gpa='3.5') select name, age from studenttab where gpa = 3.5
-PREHOOK: type: QUERY
-PREHOOK: Input: default@studenttab
-PREHOOK: Output: default@char_tbl1@gpa=3.5
-POSTHOOK: query: insert into table char_tbl1 partition(gpa='3.5') select name, age from studenttab where gpa = 3.5
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@studenttab
-POSTHOOK: Output: default@char_tbl1@gpa=3.5
-POSTHOOK: Lineage: char_tbl1 PARTITION(gpa=3.5 ).age SIMPLE [(studenttab)studenttab.FieldSchema(name:age, type:int, comment:null), ]
-POSTHOOK: Lineage: char_tbl1 PARTITION(gpa=3.5 ).name SIMPLE [(studenttab)studenttab.FieldSchema(name:name, type:string, comment:null), ]
-PREHOOK: query: insert into table char_tbl1 partition(gpa='2.5') select name, age from studenttab where gpa = 2.5
-PREHOOK: type: QUERY
-PREHOOK: Input: default@studenttab
-PREHOOK: Output: default@char_tbl1@gpa=2.5
-POSTHOOK: query: insert into table char_tbl1 partition(gpa='2.5') select name, age from studenttab where gpa = 2.5
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@studenttab
-POSTHOOK: Output: default@char_tbl1@gpa=2.5
-POSTHOOK: Lineage: char_tbl1 PARTITION(gpa=2.5 ).age SIMPLE [(studenttab)studenttab.FieldSchema(name:age, type:int, comment:null), ]
-POSTHOOK: Lineage: char_tbl1 PARTITION(gpa=2.5 ).name SIMPLE [(studenttab)studenttab.FieldSchema(name:name, type:string, comment:null), ]
-PREHOOK: query: insert into table char_tbl2 partition(gpa='3.5') select name, age from studenttab where gpa = 3.5
-PREHOOK: type: QUERY
-PREHOOK: Input: default@studenttab
-PREHOOK: Output: default@char_tbl2@gpa=3.5
-POSTHOOK: query: insert into table char_tbl2 partition(gpa='3.5') select name, age from studenttab where gpa = 3.5
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@studenttab
-POSTHOOK: Output: default@char_tbl2@gpa=3.5
-POSTHOOK: Lineage: char_tbl2 PARTITION(gpa=3.5 ).age SIMPLE [(studenttab)studenttab.FieldSchema(name:age, type:int, comment:null), ]
-POSTHOOK: Lineage: char_tbl2 PARTITION(gpa=3.5 ).name SIMPLE [(studenttab)studenttab.FieldSchema(name:name, type:string, comment:null), ]
-PREHOOK: query: insert into table char_tbl2 partition(gpa='3') select name, age from studenttab where gpa = 3
-PREHOOK: type: QUERY
-PREHOOK: Input: default@studenttab
-PREHOOK: Output: default@char_tbl2@gpa=3
-POSTHOOK: query: insert into table char_tbl2 partition(gpa='3') select name, age from studenttab where gpa = 3
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@studenttab
-POSTHOOK: Output: default@char_tbl2@gpa=3
-POSTHOOK: Lineage: char_tbl2 PARTITION(gpa=3 ).age SIMPLE [(studenttab)studenttab.FieldSchema(name:age, type:int, comment:null), ]
-POSTHOOK: Lineage: char_tbl2 PARTITION(gpa=3 ).name SIMPLE [(studenttab)studenttab.FieldSchema(name:name, type:string, comment:null), ]
-PREHOOK: query: show partitions char_tbl1
-PREHOOK: type: SHOWPARTITIONS
-PREHOOK: Input: default@char_tbl1
-POSTHOOK: query: show partitions char_tbl1
-POSTHOOK: type: SHOWPARTITIONS
-POSTHOOK: Input: default@char_tbl1
-gpa=2.5
-gpa=3.5
-PREHOOK: query: show partitions char_tbl2
-PREHOOK: type: SHOWPARTITIONS
-PREHOOK: Input: default@char_tbl2
-POSTHOOK: query: show partitions char_tbl2
-POSTHOOK: type: SHOWPARTITIONS
-POSTHOOK: Input: default@char_tbl2
-gpa=3
-gpa=3.5
-PREHOOK: query: explain vectorization select c1.name, c1.age, c1.gpa, c2.name, c2.age, c2.gpa from char_tbl1 c1 join char_tbl2 c2 on (c1.gpa = c2.gpa)
-PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization select c1.name, c1.age, c1.gpa, c2.name, c2.age, c2.gpa from char_tbl1 c1 join char_tbl2 c2 on (c1.gpa = c2.gpa)
-POSTHOOK: type: QUERY
-Plan optimized by CBO.
-
-Vertex dependency in root stage
-Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE)
-
-Stage-0
- Fetch Operator
- limit:-1
- Stage-1
- Reducer 2 llap
- File Output Operator [FS_10]
- Merge Join Operator [MERGEJOIN_21] (rows=2 width=429)
- Conds:RS_23._col2=RS_28._col2(Inner),Output:["_col0","_col1","_col2","_col3","_col4","_col5"]
- <-Map 1 [SIMPLE_EDGE] vectorized, llap
- SHUFFLE [RS_23]
- PartitionCols:_col2
- Select Operator [SEL_22] (rows=2 width=237)
- Output:["_col0","_col1","_col2"]
- TableScan [TS_0] (rows=2 width=237)
- default@char_tbl1,c1,Tbl:COMPLETE,Col:COMPLETE,Output:["name","age"]
- Dynamic Partitioning Event Operator [EVENT_26] (rows=1 width=237)
- Group By Operator [GBY_25] (rows=1 width=237)
- Output:["_col0"],keys:_col0
- Select Operator [SEL_24] (rows=2 width=237)
- Output:["_col0"]
- Please refer to the previous Select Operator [SEL_22]
- <-Map 3 [SIMPLE_EDGE] vectorized, llap
- SHUFFLE [RS_28]
- PartitionCols:_col2
- Select Operator [SEL_27] (rows=2 width=192)
- Output:["_col0","_col1","_col2"]
- TableScan [TS_3] (rows=2 width=192)
- default@char_tbl2,c2,Tbl:COMPLETE,Col:COMPLETE,Output:["name","age"]
-
-PREHOOK: query: select c1.name, c1.age, c1.gpa, c2.name, c2.age, c2.gpa from char_tbl1 c1 join char_tbl2 c2 on (c1.gpa = c2.gpa)
-PREHOOK: type: QUERY
-PREHOOK: Input: default@char_tbl1
-PREHOOK: Input: default@char_tbl1@gpa=2.5
-PREHOOK: Input: default@char_tbl1@gpa=3.5
-PREHOOK: Input: default@char_tbl2
-PREHOOK: Input: default@char_tbl2@gpa=3
-PREHOOK: Input: default@char_tbl2@gpa=3.5
-#### A masked pattern was here ####
-POSTHOOK: query: select c1.name, c1.age, c1.gpa, c2.name, c2.age, c2.gpa from char_tbl1 c1 join char_tbl2 c2 on (c1.gpa = c2.gpa)
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@char_tbl1
-POSTHOOK: Input: default@char_tbl1@gpa=2.5
-POSTHOOK: Input: default@char_tbl1@gpa=3.5
-POSTHOOK: Input: default@char_tbl2
-POSTHOOK: Input: default@char_tbl2@gpa=3
-POSTHOOK: Input: default@char_tbl2@gpa=3.5
-#### A masked pattern was here ####
-alice underhill 46 3.5 alice underhill 46 3.5
-PREHOOK: query: select c1.name, c1.age, c1.gpa, c2.name, c2.age, c2.gpa from char_tbl1 c1 join char_tbl2 c2 on (c1.gpa = c2.gpa)
-PREHOOK: type: QUERY
-PREHOOK: Input: default@char_tbl1
-PREHOOK: Input: default@char_tbl1@gpa=2.5
-PREHOOK: Input: default@char_tbl1@gpa=3.5
-PREHOOK: Input: default@char_tbl2
-PREHOOK: Input: default@char_tbl2@gpa=3
-PREHOOK: Input: default@char_tbl2@gpa=3.5
-#### A masked pattern was here ####
-POSTHOOK: query: select c1.name, c1.age, c1.gpa, c2.name, c2.age, c2.gpa from char_tbl1 c1 join char_tbl2 c2 on (c1.gpa = c2.gpa)
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@char_tbl1
-POSTHOOK: Input: default@char_tbl1@gpa=2.5
-POSTHOOK: Input: default@char_tbl1@gpa=3.5
-POSTHOOK: Input: default@char_tbl2
-POSTHOOK: Input: default@char_tbl2@gpa=3
-POSTHOOK: Input: default@char_tbl2@gpa=3.5
-#### A masked pattern was here ####
-alice underhill 46 3.5 alice underhill 46 3.5
http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/llap/vector_non_constant_in_expr.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_non_constant_in_expr.q.out b/ql/src/test/results/clientpositive/llap/vector_non_constant_in_expr.q.out
deleted file mode 100644
index d02fa08..0000000
--- a/ql/src/test/results/clientpositive/llap/vector_non_constant_in_expr.q.out
+++ /dev/null
@@ -1,51 +0,0 @@
-PREHOOK: query: explain vectorization SELECT * FROM alltypesorc WHERE cint in (ctinyint, cbigint)
-PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization SELECT * FROM alltypesorc WHERE cint in (ctinyint, cbigint)
-POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
- enabled: true
- enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
-STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
- Stage: Stage-1
- Tez
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: alltypesorc
- Statistics: Num rows: 12288 Data size: 3093170 Basic stats: COMPLETE Column stats: COMPLETE
- Filter Operator
- predicate: (cint) IN (UDFToInteger(ctinyint), UDFToInteger(cbigint)) (type: boolean)
- Statistics: Num rows: 6144 Data size: 1546640 Basic stats: COMPLETE Column stats: COMPLETE
- Select Operator
- expressions: ctinyint (type: tinyint), csmallint (type: smallint), cint (type: int), cbigint (type: bigint), cfloat (type: float), cdouble (type: double), cstring1 (type: string), cstring2 (type: string), ctimestamp1 (type: timestamp), ctimestamp2 (type: timestamp), cboolean1 (type: boolean), cboolean2 (type: boolean)
- outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11
- Statistics: Num rows: 6144 Data size: 1546640 Basic stats: COMPLETE Column stats: COMPLETE
- File Output Operator
- compressed: false
- Statistics: Num rows: 6144 Data size: 1546640 Basic stats: COMPLETE Column stats: COMPLETE
- table:
- input format: org.apache.hadoop.mapred.SequenceFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- Execution mode: llap
- LLAP IO: all inputs
- Map Vectorization:
- enabled: true
- enabledConditionsMet: hive.vectorized.use.vectorized.input.format IS true
- inputFileFormats: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
- notVectorizedReason: FILTER operator: Vectorizing IN expression only supported for constant values
- vectorized: false
-
- Stage: Stage-0
- Fetch Operator
- limit: -1
- Processor Tree:
- ListSink
-
http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/llap/vector_non_string_partition.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_non_string_partition.q.out b/ql/src/test/results/clientpositive/llap/vector_non_string_partition.q.out
deleted file mode 100644
index ee0e664..0000000
--- a/ql/src/test/results/clientpositive/llap/vector_non_string_partition.q.out
+++ /dev/null
@@ -1,274 +0,0 @@
-PREHOOK: query: CREATE TABLE non_string_part(cint INT, cstring1 STRING, cdouble DOUBLE, ctimestamp1 TIMESTAMP) PARTITIONED BY (ctinyint tinyint) STORED AS ORC
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@non_string_part
-POSTHOOK: query: CREATE TABLE non_string_part(cint INT, cstring1 STRING, cdouble DOUBLE, ctimestamp1 TIMESTAMP) PARTITIONED BY (ctinyint tinyint) STORED AS ORC
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@non_string_part
-PREHOOK: query: INSERT OVERWRITE TABLE non_string_part PARTITION(ctinyint) SELECT cint, cstring1, cdouble, ctimestamp1, ctinyint fROM alltypesorc
-WHERE ctinyint IS NULL AND cdouble IS NOT NULL ORDER BY cdouble
-PREHOOK: type: QUERY
-PREHOOK: Input: default@alltypesorc
-PREHOOK: Output: default@non_string_part
-POSTHOOK: query: INSERT OVERWRITE TABLE non_string_part PARTITION(ctinyint) SELECT cint, cstring1, cdouble, ctimestamp1, ctinyint fROM alltypesorc
-WHERE ctinyint IS NULL AND cdouble IS NOT NULL ORDER BY cdouble
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@alltypesorc
-POSTHOOK: Output: default@non_string_part@ctinyint=__HIVE_DEFAULT_PARTITION__
-POSTHOOK: Lineage: non_string_part PARTITION(ctinyint=__HIVE_DEFAULT_PARTITION__).cdouble SIMPLE [(alltypesorc)alltypesorc.FieldSchema(name:cdouble, type:double, comment:null), ]
-POSTHOOK: Lineage: non_string_part PARTITION(ctinyint=__HIVE_DEFAULT_PARTITION__).cint SIMPLE [(alltypesorc)alltypesorc.FieldSchema(name:cint, type:int, comment:null), ]
-POSTHOOK: Lineage: non_string_part PARTITION(ctinyint=__HIVE_DEFAULT_PARTITION__).cstring1 SIMPLE [(alltypesorc)alltypesorc.FieldSchema(name:cstring1, type:string, comment:null), ]
-POSTHOOK: Lineage: non_string_part PARTITION(ctinyint=__HIVE_DEFAULT_PARTITION__).ctimestamp1 SIMPLE [(alltypesorc)alltypesorc.FieldSchema(name:ctimestamp1, type:timestamp, comment:null), ]
-PREHOOK: query: SHOW PARTITIONS non_string_part
-PREHOOK: type: SHOWPARTITIONS
-PREHOOK: Input: default@non_string_part
-POSTHOOK: query: SHOW PARTITIONS non_string_part
-POSTHOOK: type: SHOWPARTITIONS
-POSTHOOK: Input: default@non_string_part
-ctinyint=__HIVE_DEFAULT_PARTITION__
-PREHOOK: query: EXPLAIN VECTORIZATION EXPRESSION SELECT cint, ctinyint FROM non_string_part WHERE cint > 0 ORDER BY cint LIMIT 10
-PREHOOK: type: QUERY
-POSTHOOK: query: EXPLAIN VECTORIZATION EXPRESSION SELECT cint, ctinyint FROM non_string_part WHERE cint > 0 ORDER BY cint LIMIT 10
-POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
- enabled: true
- enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
-STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
- Stage: Stage-1
- Tez
-#### A masked pattern was here ####
- Edges:
- Reducer 2 <- Map 1 (SIMPLE_EDGE)
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: non_string_part
- Statistics: Num rows: 3073 Data size: 24584 Basic stats: COMPLETE Column stats: COMPLETE
- TableScan Vectorization:
- native: true
- Filter Operator
- Filter Vectorization:
- className: VectorFilterOperator
- native: true
- predicateExpression: FilterLongColGreaterLongScalar(col 0:int, val 0)
- predicate: (cint > 0) (type: boolean)
- Statistics: Num rows: 1024 Data size: 8192 Basic stats: COMPLETE Column stats: COMPLETE
- Select Operator
- expressions: cint (type: int), ctinyint (type: tinyint)
- outputColumnNames: _col0, _col1
- Select Vectorization:
- className: VectorSelectOperator
- native: true
- projectedOutputColumnNums: [0, 4]
- Statistics: Num rows: 1024 Data size: 8192 Basic stats: COMPLETE Column stats: COMPLETE
- Reduce Output Operator
- key expressions: _col0 (type: int)
- sort order: +
- Reduce Sink Vectorization:
- className: VectorReduceSinkObjectHashOperator
- native: true
- nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
- Statistics: Num rows: 1024 Data size: 8192 Basic stats: COMPLETE Column stats: COMPLETE
- TopN Hash Memory Usage: 0.1
- value expressions: _col1 (type: tinyint)
- Execution mode: vectorized, llap
- LLAP IO: all inputs
- Map Vectorization:
- enabled: true
- enabledConditionsMet: hive.vectorized.use.vectorized.input.format IS true
- inputFormatFeatureSupport: []
- featureSupportInUse: []
- inputFileFormats: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
- allNative: true
- usesVectorUDFAdaptor: false
- vectorized: true
- Reducer 2
- Execution mode: vectorized, llap
- Reduce Vectorization:
- enabled: true
- enableConditionsMet: hive.vectorized.execution.reduce.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true
- allNative: false
- usesVectorUDFAdaptor: false
- vectorized: true
- Reduce Operator Tree:
- Select Operator
- expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: tinyint)
- outputColumnNames: _col0, _col1
- Select Vectorization:
- className: VectorSelectOperator
- native: true
- projectedOutputColumnNums: [0, 1]
- Statistics: Num rows: 1024 Data size: 8192 Basic stats: COMPLETE Column stats: COMPLETE
- Limit
- Number of rows: 10
- Limit Vectorization:
- className: VectorLimitOperator
- native: true
- Statistics: Num rows: 10 Data size: 80 Basic stats: COMPLETE Column stats: COMPLETE
- File Output Operator
- compressed: false
- File Sink Vectorization:
- className: VectorFileSinkOperator
- native: false
- Statistics: Num rows: 10 Data size: 80 Basic stats: COMPLETE Column stats: COMPLETE
- table:
- input format: org.apache.hadoop.mapred.SequenceFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-
- Stage: Stage-0
- Fetch Operator
- limit: 10
- Processor Tree:
- ListSink
-
-PREHOOK: query: SELECT cint, ctinyint FROM non_string_part WHERE cint > 0 ORDER BY cint LIMIT 10
-PREHOOK: type: QUERY
-PREHOOK: Input: default@non_string_part
-PREHOOK: Input: default@non_string_part@ctinyint=__HIVE_DEFAULT_PARTITION__
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT cint, ctinyint FROM non_string_part WHERE cint > 0 ORDER BY cint LIMIT 10
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@non_string_part
-POSTHOOK: Input: default@non_string_part@ctinyint=__HIVE_DEFAULT_PARTITION__
-#### A masked pattern was here ####
-762 NULL
-762 NULL
-6981 NULL
-6981 NULL
-6981 NULL
-86028 NULL
-504142 NULL
-799471 NULL
-1248059 NULL
-1286921 NULL
-PREHOOK: query: EXPLAIN VECTORIZATION EXPRESSION SELECT cint, cstring1 FROM non_string_part WHERE cint > 0 ORDER BY cint, cstring1 LIMIT 10
-PREHOOK: type: QUERY
-POSTHOOK: query: EXPLAIN VECTORIZATION EXPRESSION SELECT cint, cstring1 FROM non_string_part WHERE cint > 0 ORDER BY cint, cstring1 LIMIT 10
-POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
- enabled: true
- enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
-STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
- Stage: Stage-1
- Tez
-#### A masked pattern was here ####
- Edges:
- Reducer 2 <- Map 1 (SIMPLE_EDGE)
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: non_string_part
- Statistics: Num rows: 3073 Data size: 313446 Basic stats: COMPLETE Column stats: COMPLETE
- TableScan Vectorization:
- native: true
- Filter Operator
- Filter Vectorization:
- className: VectorFilterOperator
- native: true
- predicateExpression: FilterLongColGreaterLongScalar(col 0:int, val 0)
- predicate: (cint > 0) (type: boolean)
- Statistics: Num rows: 1024 Data size: 104448 Basic stats: COMPLETE Column stats: COMPLETE
- Select Operator
- expressions: cint (type: int), cstring1 (type: string)
- outputColumnNames: _col0, _col1
- Select Vectorization:
- className: VectorSelectOperator
- native: true
- projectedOutputColumnNums: [0, 1]
- Statistics: Num rows: 1024 Data size: 104448 Basic stats: COMPLETE Column stats: COMPLETE
- Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: string)
- sort order: ++
- Reduce Sink Vectorization:
- className: VectorReduceSinkObjectHashOperator
- native: true
- nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
- Statistics: Num rows: 1024 Data size: 104448 Basic stats: COMPLETE Column stats: COMPLETE
- TopN Hash Memory Usage: 0.1
- Execution mode: vectorized, llap
- LLAP IO: all inputs
- Map Vectorization:
- enabled: true
- enabledConditionsMet: hive.vectorized.use.vectorized.input.format IS true
- inputFormatFeatureSupport: []
- featureSupportInUse: []
- inputFileFormats: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
- allNative: true
- usesVectorUDFAdaptor: false
- vectorized: true
- Reducer 2
- Execution mode: vectorized, llap
- Reduce Vectorization:
- enabled: true
- enableConditionsMet: hive.vectorized.execution.reduce.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true
- allNative: false
- usesVectorUDFAdaptor: false
- vectorized: true
- Reduce Operator Tree:
- Select Operator
- expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: string)
- outputColumnNames: _col0, _col1
- Select Vectorization:
- className: VectorSelectOperator
- native: true
- projectedOutputColumnNums: [0, 1]
- Statistics: Num rows: 1024 Data size: 104448 Basic stats: COMPLETE Column stats: COMPLETE
- Limit
- Number of rows: 10
- Limit Vectorization:
- className: VectorLimitOperator
- native: true
- Statistics: Num rows: 10 Data size: 1020 Basic stats: COMPLETE Column stats: COMPLETE
- File Output Operator
- compressed: false
- File Sink Vectorization:
- className: VectorFileSinkOperator
- native: false
- Statistics: Num rows: 10 Data size: 1020 Basic stats: COMPLETE Column stats: COMPLETE
- table:
- input format: org.apache.hadoop.mapred.SequenceFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-
- Stage: Stage-0
- Fetch Operator
- limit: 10
- Processor Tree:
- ListSink
-
-PREHOOK: query: SELECT cint, cstring1 FROM non_string_part WHERE cint > 0 ORDER BY cint, cstring1 LIMIT 10
-PREHOOK: type: QUERY
-PREHOOK: Input: default@non_string_part
-PREHOOK: Input: default@non_string_part@ctinyint=__HIVE_DEFAULT_PARTITION__
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT cint, cstring1 FROM non_string_part WHERE cint > 0 ORDER BY cint, cstring1 LIMIT 10
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@non_string_part
-POSTHOOK: Input: default@non_string_part@ctinyint=__HIVE_DEFAULT_PARTITION__
-#### A masked pattern was here ####
-762 3WsVeqb28VWEEOLI8ail
-762 40ks5556SV
-6981 1FNNhmiFLGw425NA13g
-6981 o5mb0QP5Y48Qd4vdB0
-6981 sF2CRfgt2K
-86028 T2o8XRFAL0HC4ikDQnfoCymw
-504142 PlOxor04p5cvVl
-799471 2fu24
-1248059 Uhps6mMh3IfHB3j7yH62K
-1286921 ODLrXI8882q8LS8
http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/llap/vector_orc_string_reader_empty_dict.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_orc_string_reader_empty_dict.q.out b/ql/src/test/results/clientpositive/llap/vector_orc_string_reader_empty_dict.q.out
deleted file mode 100644
index 4f00bed..0000000
--- a/ql/src/test/results/clientpositive/llap/vector_orc_string_reader_empty_dict.q.out
+++ /dev/null
@@ -1,62 +0,0 @@
-PREHOOK: query: create table orcstr (vcol varchar(20)) stored as orc
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@orcstr
-POSTHOOK: query: create table orcstr (vcol varchar(20)) stored as orc
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@orcstr
-PREHOOK: query: insert overwrite table orcstr select null from src
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src
-PREHOOK: Output: default@orcstr
-POSTHOOK: query: insert overwrite table orcstr select null from src
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
-POSTHOOK: Output: default@orcstr
-POSTHOOK: Lineage: orcstr.vcol EXPRESSION []
-PREHOOK: query: select vcol from orcstr limit 1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@orcstr
-#### A masked pattern was here ####
-POSTHOOK: query: select vcol from orcstr limit 1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@orcstr
-#### A masked pattern was here ####
-NULL
-PREHOOK: query: select vcol from orcstr limit 1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@orcstr
-#### A masked pattern was here ####
-POSTHOOK: query: select vcol from orcstr limit 1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@orcstr
-#### A masked pattern was here ####
-NULL
-PREHOOK: query: insert overwrite table orcstr select "" from src
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src
-PREHOOK: Output: default@orcstr
-POSTHOOK: query: insert overwrite table orcstr select "" from src
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
-POSTHOOK: Output: default@orcstr
-POSTHOOK: Lineage: orcstr.vcol EXPRESSION []
-PREHOOK: query: select vcol from orcstr limit 1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@orcstr
-#### A masked pattern was here ####
-POSTHOOK: query: select vcol from orcstr limit 1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@orcstr
-#### A masked pattern was here ####
-
-PREHOOK: query: select vcol from orcstr limit 1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@orcstr
-#### A masked pattern was here ####
-POSTHOOK: query: select vcol from orcstr limit 1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@orcstr
-#### A masked pattern was here ####
-
http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/llap/vector_outer_join_no_keys.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_outer_join_no_keys.q.out b/ql/src/test/results/clientpositive/llap/vector_outer_join_no_keys.q.out
deleted file mode 100644
index 8b37aa6..0000000
--- a/ql/src/test/results/clientpositive/llap/vector_outer_join_no_keys.q.out
+++ /dev/null
@@ -1,408 +0,0 @@
-PREHOOK: query: create temporary table foo(x int) stored as orc
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@foo
-POSTHOOK: query: create temporary table foo(x int) stored as orc
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@foo
-PREHOOK: query: insert into foo values(1),(2)
-PREHOOK: type: QUERY
-PREHOOK: Input: _dummy_database@_dummy_table
-PREHOOK: Output: default@foo
-POSTHOOK: query: insert into foo values(1),(2)
-POSTHOOK: type: QUERY
-POSTHOOK: Input: _dummy_database@_dummy_table
-POSTHOOK: Output: default@foo
-POSTHOOK: Lineage: foo.x SCRIPT []
-PREHOOK: query: create temporary table bar(y int) stored as orc
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@bar
-POSTHOOK: query: create temporary table bar(y int) stored as orc
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@bar
-Warning: Map Join MAPJOIN[13][bigTable=?] in task 'Map 2' is a cross product
-PREHOOK: query: explain vectorization detail
-select count(*) from bar right outer join foo
-PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization detail
-select count(*) from bar right outer join foo
-POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
- enabled: true
- enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
-STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
- Stage: Stage-1
- Tez
-#### A masked pattern was here ####
- Edges:
- Map 2 <- Map 1 (BROADCAST_EDGE)
- Reducer 3 <- Map 2 (CUSTOM_SIMPLE_EDGE)
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: bar
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: COMPLETE
- TableScan Vectorization:
- native: true
- vectorizationSchemaColumns: [0:y:int, 1:ROW__ID:struct<transactionid:bigint,bucketid:int,rowid:bigint>]
- Select Operator
- Select Vectorization:
- className: VectorSelectOperator
- native: true
- projectedOutputColumnNums: []
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: COMPLETE
- Reduce Output Operator
- sort order:
- Reduce Sink Vectorization:
- className: VectorReduceSinkEmptyKeyOperator
- keyColumnNums: []
- native: true
- nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
- valueColumnNums: []
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: COMPLETE
- Execution mode: vectorized, llap
- LLAP IO: all inputs
- Map Vectorization:
- enabled: true
- enabledConditionsMet: hive.vectorized.use.vectorized.input.format IS true
- inputFormatFeatureSupport: []
- featureSupportInUse: []
- inputFileFormats: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
- allNative: true
- usesVectorUDFAdaptor: false
- vectorized: true
- rowBatchContext:
- dataColumnCount: 1
- includeColumns: []
- dataColumns: y:int
- partitionColumnCount: 0
- scratchColumnTypeNames: []
- Map 2
- Map Operator Tree:
- TableScan
- alias: foo
- Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
- TableScan Vectorization:
- native: true
- vectorizationSchemaColumns: [0:x:int, 1:ROW__ID:struct<transactionid:bigint,bucketid:int,rowid:bigint>]
- Select Operator
- Select Vectorization:
- className: VectorSelectOperator
- native: true
- projectedOutputColumnNums: []
- Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
- Map Join Operator
- condition map:
- Right Outer Join 0 to 1
- filter predicates:
- 0
- 1 {true}
- keys:
- 0
- 1
- Map Join Vectorization:
- className: VectorMapJoinOuterFilteredOperator
- native: false
- nativeConditionsMet: hive.mapjoin.optimized.hashtable IS true, hive.vectorized.execution.mapjoin.native.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, One MapJoin Condition IS true, No nullsafe IS true, Small table vectorizes IS true, Optimized Table and Supports Key Types IS true
- nativeConditionsNotMet: Outer Join has keys IS false
- input vertices:
- 0 Map 1
- Statistics: Num rows: 2 Data size: 10 Basic stats: PARTIAL Column stats: NONE
- Group By Operator
- aggregations: count()
- Group By Vectorization:
- aggregators: VectorUDAFCountStar(*) -> bigint
- className: VectorGroupByOperator
- groupByMode: HASH
- native: false
- vectorProcessingMode: HASH
- projectedOutputColumnNums: [0]
- mode: hash
- outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE
- Reduce Output Operator
- sort order:
- Reduce Sink Vectorization:
- className: VectorReduceSinkEmptyKeyOperator
- keyColumnNums: []
- native: true
- nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
- valueColumnNums: [0]
- Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE
- value expressions: _col0 (type: bigint)
- Execution mode: vectorized, llap
- LLAP IO: all inputs
- Map Vectorization:
- enabled: true
- enabledConditionsMet: hive.vectorized.use.vectorized.input.format IS true
- inputFormatFeatureSupport: []
- featureSupportInUse: []
- inputFileFormats: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
- allNative: false
- usesVectorUDFAdaptor: false
- vectorized: true
- rowBatchContext:
- dataColumnCount: 1
- includeColumns: []
- dataColumns: x:int
- partitionColumnCount: 0
- scratchColumnTypeNames: []
- Reducer 3
- Execution mode: vectorized, llap
- Reduce Vectorization:
- enabled: true
- enableConditionsMet: hive.vectorized.execution.reduce.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true
- reduceColumnNullOrder:
- reduceColumnSortOrder:
- allNative: false
- usesVectorUDFAdaptor: false
- vectorized: true
- rowBatchContext:
- dataColumnCount: 1
- dataColumns: VALUE._col0:bigint
- partitionColumnCount: 0
- scratchColumnTypeNames: []
- Reduce Operator Tree:
- Group By Operator
- aggregations: count(VALUE._col0)
- Group By Vectorization:
- aggregators: VectorUDAFCountMerge(col 0:bigint) -> bigint
- className: VectorGroupByOperator
- groupByMode: MERGEPARTIAL
- native: false
- vectorProcessingMode: GLOBAL
- projectedOutputColumnNums: [0]
- mode: mergepartial
- outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE
- File Output Operator
- compressed: false
- File Sink Vectorization:
- className: VectorFileSinkOperator
- native: false
- Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.SequenceFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-
- Stage: Stage-0
- Fetch Operator
- limit: -1
- Processor Tree:
- ListSink
-
-Warning: Map Join MAPJOIN[13][bigTable=?] in task 'Map 2' is a cross product
-PREHOOK: query: -- = 2
-
-select count(*) from bar right outer join foo
-PREHOOK: type: QUERY
-PREHOOK: Input: default@bar
-PREHOOK: Input: default@foo
-#### A masked pattern was here ####
-POSTHOOK: query: -- = 2
-
-select count(*) from bar right outer join foo
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@bar
-POSTHOOK: Input: default@foo
-#### A masked pattern was here ####
-2
-Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product
-PREHOOK: query: -- = 2
-
-explain vectorization detail
-select count(*) from bar, foo
-PREHOOK: type: QUERY
-POSTHOOK: query: -- = 2
-
-explain vectorization detail
-select count(*) from bar, foo
-POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
- enabled: true
- enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
-STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
- Stage: Stage-1
- Tez
-#### A masked pattern was here ####
- Edges:
- Reducer 2 <- Map 1 (XPROD_EDGE), Map 4 (XPROD_EDGE)
- Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE)
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: bar
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: COMPLETE
- TableScan Vectorization:
- native: true
- vectorizationSchemaColumns: [0:y:int, 1:ROW__ID:struct<transactionid:bigint,bucketid:int,rowid:bigint>]
- Select Operator
- Select Vectorization:
- className: VectorSelectOperator
- native: true
- projectedOutputColumnNums: []
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: COMPLETE
- Reduce Output Operator
- sort order:
- Reduce Sink Vectorization:
- className: VectorReduceSinkEmptyKeyOperator
- keyColumnNums: []
- native: true
- nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
- valueColumnNums: []
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: COMPLETE
- Execution mode: vectorized, llap
- LLAP IO: all inputs
- Map Vectorization:
- enabled: true
- enabledConditionsMet: hive.vectorized.use.vectorized.input.format IS true
- inputFormatFeatureSupport: []
- featureSupportInUse: []
- inputFileFormats: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
- allNative: true
- usesVectorUDFAdaptor: false
- vectorized: true
- rowBatchContext:
- dataColumnCount: 1
- includeColumns: []
- dataColumns: y:int
- partitionColumnCount: 0
- scratchColumnTypeNames: []
- Map 4
- Map Operator Tree:
- TableScan
- alias: foo
- Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
- TableScan Vectorization:
- native: true
- vectorizationSchemaColumns: [0:x:int, 1:ROW__ID:struct<transactionid:bigint,bucketid:int,rowid:bigint>]
- Select Operator
- Select Vectorization:
- className: VectorSelectOperator
- native: true
- projectedOutputColumnNums: []
- Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
- Reduce Output Operator
- sort order:
- Reduce Sink Vectorization:
- className: VectorReduceSinkEmptyKeyOperator
- keyColumnNums: []
- native: true
- nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
- valueColumnNums: []
- Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
- Execution mode: vectorized, llap
- LLAP IO: all inputs
- Map Vectorization:
- enabled: true
- enabledConditionsMet: hive.vectorized.use.vectorized.input.format IS true
- inputFormatFeatureSupport: []
- featureSupportInUse: []
- inputFileFormats: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
- allNative: true
- usesVectorUDFAdaptor: false
- vectorized: true
- rowBatchContext:
- dataColumnCount: 1
- includeColumns: []
- dataColumns: x:int
- partitionColumnCount: 0
- scratchColumnTypeNames: []
- Reducer 2
- Execution mode: llap
- Reduce Operator Tree:
- Merge Join Operator
- condition map:
- Inner Join 0 to 1
- keys:
- 0
- 1
- Statistics: Num rows: 2 Data size: 10 Basic stats: PARTIAL Column stats: NONE
- Group By Operator
- aggregations: count()
- mode: hash
- outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE
- Reduce Output Operator
- sort order:
- Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE
- value expressions: _col0 (type: bigint)
- Reducer 3
- Execution mode: vectorized, llap
- Reduce Vectorization:
- enabled: true
- enableConditionsMet: hive.vectorized.execution.reduce.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true
- reduceColumnNullOrder:
- reduceColumnSortOrder:
- allNative: false
- usesVectorUDFAdaptor: false
- vectorized: true
- rowBatchContext:
- dataColumnCount: 1
- dataColumns: VALUE._col0:bigint
- partitionColumnCount: 0
- scratchColumnTypeNames: []
- Reduce Operator Tree:
- Group By Operator
- aggregations: count(VALUE._col0)
- Group By Vectorization:
- aggregators: VectorUDAFCountMerge(col 0:bigint) -> bigint
- className: VectorGroupByOperator
- groupByMode: MERGEPARTIAL
- native: false
- vectorProcessingMode: GLOBAL
- projectedOutputColumnNums: [0]
- mode: mergepartial
- outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE
- File Output Operator
- compressed: false
- File Sink Vectorization:
- className: VectorFileSinkOperator
- native: false
- Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.SequenceFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-
- Stage: Stage-0
- Fetch Operator
- limit: -1
- Processor Tree:
- ListSink
-
-Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product
-PREHOOK: query: -- = 0
-
-select count(*) from bar, foo
-PREHOOK: type: QUERY
-PREHOOK: Input: default@bar
-PREHOOK: Input: default@foo
-#### A masked pattern was here ####
-POSTHOOK: query: -- = 0
-
-select count(*) from bar, foo
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@bar
-POSTHOOK: Input: default@foo
-#### A masked pattern was here ####
-0
http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/llap/vector_tablesample_rows.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_tablesample_rows.q.out b/ql/src/test/results/clientpositive/llap/vector_tablesample_rows.q.out
deleted file mode 100644
index 4429e9a..0000000
--- a/ql/src/test/results/clientpositive/llap/vector_tablesample_rows.q.out
+++ /dev/null
@@ -1,400 +0,0 @@
-PREHOOK: query: explain vectorization detail
-select 'key1', 'value1' from alltypesorc tablesample (1 rows)
-PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization detail
-select 'key1', 'value1' from alltypesorc tablesample (1 rows)
-POSTHOOK: type: QUERY
-Explain
-PLAN VECTORIZATION:
- enabled: true
- enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
-STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
- Stage: Stage-1
- Tez
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: alltypesorc
- Row Limit Per Split: 1
- Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: COMPLETE
- TableScan Vectorization:
- native: true
- vectorizationSchemaColumns: [0:ctinyint:tinyint, 1:csmallint:smallint, 2:cint:int, 3:cbigint:bigint, 4:cfloat:float, 5:cdouble:double, 6:cstring1:string, 7:cstring2:string, 8:ctimestamp1:timestamp, 9:ctimestamp2:timestamp, 10:cboolean1:boolean, 11:cboolean2:boolean, 12:ROW__ID:struct<transactionid:bigint,bucketid:int,rowid:bigint>]
- Select Operator
- expressions: 'key1' (type: string), 'value1' (type: string)
- outputColumnNames: _col0, _col1
- Select Vectorization:
- className: VectorSelectOperator
- native: true
- projectedOutputColumnNums: [13, 14]
- selectExpressions: ConstantVectorExpression(val key1) -> 13:string, ConstantVectorExpression(val value1) -> 14:string
- Statistics: Num rows: 12288 Data size: 2187264 Basic stats: COMPLETE Column stats: COMPLETE
- File Output Operator
- compressed: false
- File Sink Vectorization:
- className: VectorFileSinkOperator
- native: false
- Statistics: Num rows: 12288 Data size: 2187264 Basic stats: COMPLETE Column stats: COMPLETE
- table:
- input format: org.apache.hadoop.mapred.SequenceFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- Execution mode: vectorized, llap
- LLAP IO: all inputs
- Map Vectorization:
- enabled: true
- enabledConditionsMet: hive.vectorized.use.vectorized.input.format IS true
- inputFormatFeatureSupport: []
- featureSupportInUse: []
- inputFileFormats: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
- allNative: false
- usesVectorUDFAdaptor: false
- vectorized: true
- rowBatchContext:
- dataColumnCount: 12
- includeColumns: []
- dataColumns: ctinyint:tinyint, csmallint:smallint, cint:int, cbigint:bigint, cfloat:float, cdouble:double, cstring1:string, cstring2:string, ctimestamp1:timestamp, ctimestamp2:timestamp, cboolean1:boolean, cboolean2:boolean
- partitionColumnCount: 0
- scratchColumnTypeNames: [string, string]
-
- Stage: Stage-0
- Fetch Operator
- limit: -1
- Processor Tree:
- ListSink
-
-PREHOOK: query: select 'key1', 'value1' from alltypesorc tablesample (1 rows)
-PREHOOK: type: QUERY
-PREHOOK: Input: default@alltypesorc
-#### A masked pattern was here ####
-POSTHOOK: query: select 'key1', 'value1' from alltypesorc tablesample (1 rows)
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@alltypesorc
-#### A masked pattern was here ####
-_c0 _c1
-key1 value1
-PREHOOK: query: create table decimal_2 (t decimal(18,9)) stored as orc
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@decimal_2
-POSTHOOK: query: create table decimal_2 (t decimal(18,9)) stored as orc
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@decimal_2
-PREHOOK: query: explain vectorization detail
-insert overwrite table decimal_2
- select cast('17.29' as decimal(4,2)) from alltypesorc tablesample (1 rows)
-PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization detail
-insert overwrite table decimal_2
- select cast('17.29' as decimal(4,2)) from alltypesorc tablesample (1 rows)
-POSTHOOK: type: QUERY
-Explain
-PLAN VECTORIZATION:
- enabled: true
- enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
-STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-2 depends on stages: Stage-1
- Stage-0 depends on stages: Stage-2
- Stage-3 depends on stages: Stage-0
-
-STAGE PLANS:
- Stage: Stage-1
- Tez
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: alltypesorc
- Row Limit Per Split: 1
- Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: COMPLETE
- TableScan Vectorization:
- native: true
- vectorizationSchemaColumns: [0:ctinyint:tinyint, 1:csmallint:smallint, 2:cint:int, 3:cbigint:bigint, 4:cfloat:float, 5:cdouble:double, 6:cstring1:string, 7:cstring2:string, 8:ctimestamp1:timestamp, 9:ctimestamp2:timestamp, 10:cboolean1:boolean, 11:cboolean2:boolean, 12:ROW__ID:struct<transactionid:bigint,bucketid:int,rowid:bigint>]
- Select Operator
- expressions: 17.29 (type: decimal(18,9))
- outputColumnNames: _col0
- Select Vectorization:
- className: VectorSelectOperator
- native: true
- projectedOutputColumnNums: [13]
- selectExpressions: ConstantVectorExpression(val 17.29) -> 13:decimal(18,9)
- Statistics: Num rows: 12288 Data size: 1376256 Basic stats: COMPLETE Column stats: COMPLETE
- File Output Operator
- compressed: false
- File Sink Vectorization:
- className: VectorFileSinkOperator
- native: false
- Statistics: Num rows: 12288 Data size: 1376256 Basic stats: COMPLETE Column stats: COMPLETE
- table:
- input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
- output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
- serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
- name: default.decimal_2
- Execution mode: vectorized, llap
- LLAP IO: all inputs
- Map Vectorization:
- enabled: true
- enabledConditionsMet: hive.vectorized.use.vectorized.input.format IS true
- inputFormatFeatureSupport: []
- featureSupportInUse: []
- inputFileFormats: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
- allNative: false
- usesVectorUDFAdaptor: false
- vectorized: true
- rowBatchContext:
- dataColumnCount: 12
- includeColumns: []
- dataColumns: ctinyint:tinyint, csmallint:smallint, cint:int, cbigint:bigint, cfloat:float, cdouble:double, cstring1:string, cstring2:string, ctimestamp1:timestamp, ctimestamp2:timestamp, cboolean1:boolean, cboolean2:boolean
- partitionColumnCount: 0
- scratchColumnTypeNames: [decimal(18,9)]
-
- Stage: Stage-2
- Dependency Collection
-
- Stage: Stage-0
- Move Operator
- tables:
- replace: true
- table:
- input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
- output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
- serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
- name: default.decimal_2
-
- Stage: Stage-3
- Stats Work
- Basic Stats Work:
-
-PREHOOK: query: insert overwrite table decimal_2
- select cast('17.29' as decimal(4,2)) from alltypesorc tablesample (1 rows)
-PREHOOK: type: QUERY
-PREHOOK: Input: default@alltypesorc
-PREHOOK: Output: default@decimal_2
-POSTHOOK: query: insert overwrite table decimal_2
- select cast('17.29' as decimal(4,2)) from alltypesorc tablesample (1 rows)
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@alltypesorc
-POSTHOOK: Output: default@decimal_2
-POSTHOOK: Lineage: decimal_2.t EXPRESSION []
-_col0
-PREHOOK: query: select count(*) from decimal_2
-PREHOOK: type: QUERY
-PREHOOK: Input: default@decimal_2
-#### A masked pattern was here ####
-POSTHOOK: query: select count(*) from decimal_2
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@decimal_2
-#### A masked pattern was here ####
-_c0
-1
-PREHOOK: query: drop table decimal_2
-PREHOOK: type: DROPTABLE
-PREHOOK: Input: default@decimal_2
-PREHOOK: Output: default@decimal_2
-POSTHOOK: query: drop table decimal_2
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Input: default@decimal_2
-POSTHOOK: Output: default@decimal_2
-PREHOOK: query: explain vectorization detail
-select count(1) from (select * from (Select 1 a) x order by x.a) y
-PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization detail
-select count(1) from (select * from (Select 1 a) x order by x.a) y
-POSTHOOK: type: QUERY
-Explain
-PLAN VECTORIZATION:
- enabled: true
- enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
-STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
- Stage: Stage-1
- Tez
-#### A masked pattern was here ####
- Edges:
- Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE)
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: _dummy_table
- Row Limit Per Split: 1
- Statistics: Num rows: 1 Data size: 10 Basic stats: COMPLETE Column stats: COMPLETE
- Select Operator
- Statistics: Num rows: 1 Data size: 10 Basic stats: COMPLETE Column stats: COMPLETE
- Group By Operator
- aggregations: count()
- mode: hash
- outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
- Reduce Output Operator
- sort order:
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
- value expressions: _col0 (type: bigint)
- Execution mode: llap
- LLAP IO: no inputs
- Map Vectorization:
- enabled: false
-#### A masked pattern was here ####
- Reducer 2
- Execution mode: vectorized, llap
- Reduce Vectorization:
- enabled: true
- enableConditionsMet: hive.vectorized.execution.reduce.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true
- reduceColumnNullOrder:
- reduceColumnSortOrder:
- allNative: false
- usesVectorUDFAdaptor: false
- vectorized: true
- rowBatchContext:
- dataColumnCount: 1
- dataColumns: VALUE._col0:bigint
- partitionColumnCount: 0
- scratchColumnTypeNames: []
- Reduce Operator Tree:
- Group By Operator
- aggregations: count(VALUE._col0)
- Group By Vectorization:
- aggregators: VectorUDAFCountMerge(col 0:bigint) -> bigint
- className: VectorGroupByOperator
- groupByMode: MERGEPARTIAL
- native: false
- vectorProcessingMode: GLOBAL
- projectedOutputColumnNums: [0]
- mode: mergepartial
- outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
- File Output Operator
- compressed: false
- File Sink Vectorization:
- className: VectorFileSinkOperator
- native: false
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
- table:
- input format: org.apache.hadoop.mapred.SequenceFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-
- Stage: Stage-0
- Fetch Operator
- limit: -1
- Processor Tree:
- ListSink
-
-PREHOOK: query: select count(1) from (select * from (Select 1 a) x order by x.a) y
-PREHOOK: type: QUERY
-PREHOOK: Input: _dummy_database@_dummy_table
-#### A masked pattern was here ####
-POSTHOOK: query: select count(1) from (select * from (Select 1 a) x order by x.a) y
-POSTHOOK: type: QUERY
-POSTHOOK: Input: _dummy_database@_dummy_table
-#### A masked pattern was here ####
-_c0
-1
-PREHOOK: query: explain vectorization detail
-create temporary table dual as select 1
-PREHOOK: type: CREATETABLE_AS_SELECT
-POSTHOOK: query: explain vectorization detail
-create temporary table dual as select 1
-POSTHOOK: type: CREATETABLE_AS_SELECT
-Explain
-PLAN VECTORIZATION:
- enabled: true
- enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
-STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-2 depends on stages: Stage-1
- Stage-4 depends on stages: Stage-2, Stage-0
- Stage-3 depends on stages: Stage-4
- Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
- Stage: Stage-1
- Tez
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: _dummy_table
- Row Limit Per Split: 1
- Statistics: Num rows: 1 Data size: 10 Basic stats: COMPLETE Column stats: COMPLETE
- Select Operator
- expressions: 1 (type: int)
- outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE
- File Output Operator
- compressed: false
- Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.dual
- Execution mode: llap
- LLAP IO: no inputs
- Map Vectorization:
- enabled: false
-#### A masked pattern was here ####
-
- Stage: Stage-2
- Dependency Collection
-
- Stage: Stage-4
- Create Table Operator:
- Create Table
- columns: _c0 int
- input format: org.apache.hadoop.mapred.TextInputFormat
-#### A masked pattern was here ####
- output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
- serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.dual
- isTemporary: true
-
- Stage: Stage-3
- Stats Work
- Basic Stats Work:
-
- Stage: Stage-0
- Move Operator
- files:
- hdfs directory: true
-#### A masked pattern was here ####
-
-PREHOOK: query: create temporary table dual as select 1
-PREHOOK: type: CREATETABLE_AS_SELECT
-PREHOOK: Input: _dummy_database@_dummy_table
-PREHOOK: Output: database:default
-PREHOOK: Output: default@dual
-POSTHOOK: query: create temporary table dual as select 1
-POSTHOOK: type: CREATETABLE_AS_SELECT
-POSTHOOK: Input: _dummy_database@_dummy_table
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@dual
-_c0
-PREHOOK: query: select * from dual
-PREHOOK: type: QUERY
-PREHOOK: Input: default@dual
-#### A masked pattern was here ####
-POSTHOOK: query: select * from dual
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@dual
-#### A masked pattern was here ####
-dual._c0
-1