You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by se...@apache.org on 2018/07/14 01:53:40 UTC
[18/20] hive git commit: HIVE-19820 : add ACID stats support to
background stats updater and fix bunch of edge cases found in SU tests
(Sergey Shelukhin)
http://git-wip-us.apache.org/repos/asf/hive/blob/1c9947f3/ql/src/test/results/clientpositive/llap/results_cache_truncate.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/results_cache_truncate.q.out b/ql/src/test/results/clientpositive/llap/results_cache_truncate.q.out
index 0b7a81a..83b2144 100644
--- a/ql/src/test/results/clientpositive/llap/results_cache_truncate.q.out
+++ b/ql/src/test/results/clientpositive/llap/results_cache_truncate.q.out
@@ -23,54 +23,12 @@ POSTHOOK: query: explain
select count(*) from rct1_1
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-0 depends on stages: Stage-1
+ Stage-0 is a root stage
STAGE PLANS:
- Stage: Stage-1
- Tez
-#### A masked pattern was here ####
- Edges:
- Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE)
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: rct1_1
- Statistics: Num rows: 500 Data size: 35250 Basic stats: COMPLETE Column stats: COMPLETE
- Select Operator
- Statistics: Num rows: 500 Data size: 35250 Basic stats: COMPLETE Column stats: COMPLETE
- Group By Operator
- aggregations: count()
- mode: hash
- outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
- Reduce Output Operator
- sort order:
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
- value expressions: _col0 (type: bigint)
- Execution mode: vectorized, llap
- LLAP IO: may be used (ACID table)
- Reducer 2
- Execution mode: vectorized, llap
- Reduce Operator Tree:
- Group By Operator
- aggregations: count(VALUE._col0)
- mode: mergepartial
- outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
- File Output Operator
- compressed: false
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
- table:
- input format: org.apache.hadoop.mapred.SequenceFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-
Stage: Stage-0
Fetch Operator
- limit: -1
+ limit: 1
Processor Tree:
ListSink
@@ -96,17 +54,18 @@ STAGE DEPENDENCIES:
STAGE PLANS:
Stage: Stage-0
Fetch Operator
- limit: -1
+ limit: 1
Processor Tree:
ListSink
- Cached Query Result: true
PREHOOK: query: select count(*) from rct1_1
PREHOOK: type: QUERY
PREHOOK: Input: default@rct1_1
+#### A masked pattern was here ####
POSTHOOK: query: select count(*) from rct1_1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@rct1_1
+#### A masked pattern was here ####
500
PREHOOK: query: truncate table rct1_1
PREHOOK: type: TRUNCATETABLE
@@ -122,54 +81,12 @@ POSTHOOK: query: explain
select count(*) from rct1_1
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-0 depends on stages: Stage-1
+ Stage-0 is a root stage
STAGE PLANS:
- Stage: Stage-1
- Tez
-#### A masked pattern was here ####
- Edges:
- Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE)
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: rct1_1
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: COMPLETE
- Select Operator
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: COMPLETE
- Group By Operator
- aggregations: count()
- mode: hash
- outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: COMPLETE
- Reduce Output Operator
- sort order:
- Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: COMPLETE
- value expressions: _col0 (type: bigint)
- Execution mode: vectorized, llap
- LLAP IO: may be used (ACID table)
- Reducer 2
- Execution mode: vectorized, llap
- Reduce Operator Tree:
- Group By Operator
- aggregations: count(VALUE._col0)
- mode: mergepartial
- outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: COMPLETE
- File Output Operator
- compressed: false
- Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: COMPLETE
- table:
- input format: org.apache.hadoop.mapred.SequenceFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-
Stage: Stage-0
Fetch Operator
- limit: -1
+ limit: 1
Processor Tree:
ListSink
@@ -217,68 +134,22 @@ POSTHOOK: query: explain
select count(*) from rct1_2
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-0 depends on stages: Stage-1
+ Stage-0 is a root stage
STAGE PLANS:
- Stage: Stage-1
- Tez
-#### A masked pattern was here ####
- Edges:
- Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE)
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: rct1_2
- Statistics: Num rows: 1000 Data size: 78450 Basic stats: COMPLETE Column stats: COMPLETE
- Select Operator
- Statistics: Num rows: 1000 Data size: 78450 Basic stats: COMPLETE Column stats: COMPLETE
- Group By Operator
- aggregations: count()
- mode: hash
- outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
- Reduce Output Operator
- sort order:
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
- value expressions: _col0 (type: bigint)
- Execution mode: vectorized, llap
- LLAP IO: may be used (ACID table)
- Reducer 2
- Execution mode: vectorized, llap
- Reduce Operator Tree:
- Group By Operator
- aggregations: count(VALUE._col0)
- mode: mergepartial
- outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
- File Output Operator
- compressed: false
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
- table:
- input format: org.apache.hadoop.mapred.SequenceFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-
Stage: Stage-0
Fetch Operator
- limit: -1
+ limit: 1
Processor Tree:
ListSink
PREHOOK: query: select count(*) from rct1_2
PREHOOK: type: QUERY
PREHOOK: Input: default@rct1_2
-PREHOOK: Input: default@rct1_2@p1=part1
-PREHOOK: Input: default@rct1_2@p1=part2
#### A masked pattern was here ####
POSTHOOK: query: select count(*) from rct1_2
POSTHOOK: type: QUERY
POSTHOOK: Input: default@rct1_2
-POSTHOOK: Input: default@rct1_2@p1=part1
-POSTHOOK: Input: default@rct1_2@p1=part2
#### A masked pattern was here ####
1000
test.comment="Query on transactional table should use cache"
@@ -294,21 +165,18 @@ STAGE DEPENDENCIES:
STAGE PLANS:
Stage: Stage-0
Fetch Operator
- limit: -1
+ limit: 1
Processor Tree:
ListSink
- Cached Query Result: true
PREHOOK: query: select count(*) from rct1_2
PREHOOK: type: QUERY
PREHOOK: Input: default@rct1_2
-PREHOOK: Input: default@rct1_2@p1=part1
-PREHOOK: Input: default@rct1_2@p1=part2
+#### A masked pattern was here ####
POSTHOOK: query: select count(*) from rct1_2
POSTHOOK: type: QUERY
POSTHOOK: Input: default@rct1_2
-POSTHOOK: Input: default@rct1_2@p1=part1
-POSTHOOK: Input: default@rct1_2@p1=part2
+#### A masked pattern was here ####
1000
PREHOOK: query: truncate table rct1_2 partition (p1='part1')
PREHOOK: type: TRUNCATETABLE
@@ -324,68 +192,22 @@ POSTHOOK: query: explain
select count(*) from rct1_2
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-0 depends on stages: Stage-1
+ Stage-0 is a root stage
STAGE PLANS:
- Stage: Stage-1
- Tez
-#### A masked pattern was here ####
- Edges:
- Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE)
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: rct1_2
- Statistics: Num rows: 500 Data size: 39200 Basic stats: PARTIAL Column stats: COMPLETE
- Select Operator
- Statistics: Num rows: 500 Data size: 39200 Basic stats: PARTIAL Column stats: COMPLETE
- Group By Operator
- aggregations: count()
- mode: hash
- outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: COMPLETE
- Reduce Output Operator
- sort order:
- Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: COMPLETE
- value expressions: _col0 (type: bigint)
- Execution mode: vectorized, llap
- LLAP IO: may be used (ACID table)
- Reducer 2
- Execution mode: vectorized, llap
- Reduce Operator Tree:
- Group By Operator
- aggregations: count(VALUE._col0)
- mode: mergepartial
- outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: COMPLETE
- File Output Operator
- compressed: false
- Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: COMPLETE
- table:
- input format: org.apache.hadoop.mapred.SequenceFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-
Stage: Stage-0
Fetch Operator
- limit: -1
+ limit: 1
Processor Tree:
ListSink
PREHOOK: query: select count(*) from rct1_2
PREHOOK: type: QUERY
PREHOOK: Input: default@rct1_2
-PREHOOK: Input: default@rct1_2@p1=part1
-PREHOOK: Input: default@rct1_2@p1=part2
#### A masked pattern was here ####
POSTHOOK: query: select count(*) from rct1_2
POSTHOOK: type: QUERY
POSTHOOK: Input: default@rct1_2
-POSTHOOK: Input: default@rct1_2@p1=part1
-POSTHOOK: Input: default@rct1_2@p1=part2
#### A masked pattern was here ####
500
PREHOOK: query: truncate table rct1_2
@@ -404,67 +226,21 @@ POSTHOOK: query: explain
select count(*) from rct1_2
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-1 is a root stage
- Stage-0 depends on stages: Stage-1
+ Stage-0 is a root stage
STAGE PLANS:
- Stage: Stage-1
- Tez
-#### A masked pattern was here ####
- Edges:
- Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE)
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: rct1_2
- Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: COMPLETE
- Select Operator
- Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: COMPLETE
- Group By Operator
- aggregations: count()
- mode: hash
- outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: COMPLETE
- Reduce Output Operator
- sort order:
- Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: COMPLETE
- value expressions: _col0 (type: bigint)
- Execution mode: vectorized, llap
- LLAP IO: may be used (ACID table)
- Reducer 2
- Execution mode: vectorized, llap
- Reduce Operator Tree:
- Group By Operator
- aggregations: count(VALUE._col0)
- mode: mergepartial
- outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: COMPLETE
- File Output Operator
- compressed: false
- Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: COMPLETE
- table:
- input format: org.apache.hadoop.mapred.SequenceFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-
Stage: Stage-0
Fetch Operator
- limit: -1
+ limit: 1
Processor Tree:
ListSink
PREHOOK: query: select count(*) from rct1_2
PREHOOK: type: QUERY
PREHOOK: Input: default@rct1_2
-PREHOOK: Input: default@rct1_2@p1=part1
-PREHOOK: Input: default@rct1_2@p1=part2
#### A masked pattern was here ####
POSTHOOK: query: select count(*) from rct1_2
POSTHOOK: type: QUERY
POSTHOOK: Input: default@rct1_2
-POSTHOOK: Input: default@rct1_2@p1=part1
-POSTHOOK: Input: default@rct1_2@p1=part2
#### A masked pattern was here ####
0
http://git-wip-us.apache.org/repos/asf/hive/blob/1c9947f3/ql/src/test/results/clientpositive/llap/sqlmerge_stats.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/sqlmerge_stats.q.out b/ql/src/test/results/clientpositive/llap/sqlmerge_stats.q.out
index 94d0ea3..eba16aa 100644
--- a/ql/src/test/results/clientpositive/llap/sqlmerge_stats.q.out
+++ b/ql/src/test/results/clientpositive/llap/sqlmerge_stats.q.out
@@ -88,6 +88,7 @@ Retention: 0
#### A masked pattern was here ####
Table Type: MANAGED_TABLE
Table Parameters:
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"a\":\"true\",\"b\":\"true\"}}
bucketing_version 2
numFiles 1
numRows 1
@@ -141,12 +142,12 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: t
- Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
key expressions: a (type: int)
sort order: +
Map-reduce partition columns: a (type: int)
- Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE
value expressions: ROW__ID (type: struct<writeid:bigint,bucketid:int,rowid:bigint>)
Execution mode: vectorized, llap
LLAP IO: may be used (ACID table)
@@ -173,50 +174,50 @@ STAGE PLANS:
0 a (type: int)
1 a (type: int)
outputColumnNames: _col0, _col4, _col5, _col6
- Statistics: Num rows: 2 Data size: 17 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 2 Data size: 176 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
predicate: (_col0 = _col5) (type: boolean)
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 88 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col4 (type: struct<writeid:bigint,bucketid:int,rowid:bigint>), _col0 (type: int)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
key expressions: _col0 (type: struct<writeid:bigint,bucketid:int,rowid:bigint>)
sort order: +
Map-reduce partition columns: UDFToInteger(_col0) (type: int)
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE
value expressions: _col1 (type: int)
Filter Operator
predicate: (_col0 = _col5) (type: boolean)
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 88 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col4 (type: struct<writeid:bigint,bucketid:int,rowid:bigint>)
outputColumnNames: _col4
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 88 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: count()
keys: _col4 (type: struct<writeid:bigint,bucketid:int,rowid:bigint>)
mode: hash
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
key expressions: _col0 (type: struct<writeid:bigint,bucketid:int,rowid:bigint>)
sort order: +
Map-reduce partition columns: _col0 (type: struct<writeid:bigint,bucketid:int,rowid:bigint>)
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE
value expressions: _col1 (type: bigint)
Filter Operator
predicate: _col0 is null (type: boolean)
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 88 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col5 (type: int), _col6 (type: int)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
sort order:
Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
value expressions: _col0 (type: int), _col1 (type: int)
Reducer 3
Execution mode: vectorized, llap
@@ -224,10 +225,10 @@ STAGE PLANS:
Select Operator
expressions: KEY.reducesinkkey0 (type: struct<writeid:bigint,bucketid:int,rowid:bigint>), VALUE._col0 (type: int), 99 (type: int)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
@@ -242,17 +243,17 @@ STAGE PLANS:
keys: KEY._col0 (type: struct<writeid:bigint,bucketid:int,rowid:bigint>)
mode: mergepartial
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
predicate: (_col1 > 1L) (type: boolean)
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: cardinality_violation(_col0) (type: int)
outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -261,19 +262,19 @@ STAGE PLANS:
Select Operator
expressions: _col0 (type: int)
outputColumnNames: val
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: compute_stats(val, 'hll')
mode: complete
outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 432 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col0 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>)
outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 432 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 432 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -284,10 +285,10 @@ STAGE PLANS:
Select Operator
expressions: VALUE._col0 (type: int), VALUE._col1 (type: int)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
@@ -297,15 +298,15 @@ STAGE PLANS:
Select Operator
expressions: _col0 (type: int), _col1 (type: int)
outputColumnNames: a, b
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: compute_stats(a, 'hll'), compute_stats(b, 'hll')
mode: hash
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
sort order:
- Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE
value expressions: _col0 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>), _col1 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>)
Reducer 6
Execution mode: llap
@@ -314,10 +315,10 @@ STAGE PLANS:
aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1)
mode: mergepartial
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -425,6 +426,7 @@ Retention: 0
#### A masked pattern was here ####
Table Type: MANAGED_TABLE
Table Parameters:
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
bucketing_version 2
numFiles 4
numRows 2
@@ -490,6 +492,7 @@ Retention: 0
#### A masked pattern was here ####
Table Type: MANAGED_TABLE
Table Parameters:
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
bucketing_version 2
numFiles 6
numRows 0
http://git-wip-us.apache.org/repos/asf/hive/blob/1c9947f3/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AbortTxnsRequest.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AbortTxnsRequest.java b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AbortTxnsRequest.java
index 3ee3370..7dcfc17 100644
--- a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AbortTxnsRequest.java
+++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AbortTxnsRequest.java
@@ -351,13 +351,13 @@ import org.slf4j.LoggerFactory;
case 1: // TXN_IDS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list586 = iprot.readListBegin();
- struct.txn_ids = new ArrayList<Long>(_list586.size);
- long _elem587;
- for (int _i588 = 0; _i588 < _list586.size; ++_i588)
+ org.apache.thrift.protocol.TList _list594 = iprot.readListBegin();
+ struct.txn_ids = new ArrayList<Long>(_list594.size);
+ long _elem595;
+ for (int _i596 = 0; _i596 < _list594.size; ++_i596)
{
- _elem587 = iprot.readI64();
- struct.txn_ids.add(_elem587);
+ _elem595 = iprot.readI64();
+ struct.txn_ids.add(_elem595);
}
iprot.readListEnd();
}
@@ -383,9 +383,9 @@ import org.slf4j.LoggerFactory;
oprot.writeFieldBegin(TXN_IDS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, struct.txn_ids.size()));
- for (long _iter589 : struct.txn_ids)
+ for (long _iter597 : struct.txn_ids)
{
- oprot.writeI64(_iter589);
+ oprot.writeI64(_iter597);
}
oprot.writeListEnd();
}
@@ -410,9 +410,9 @@ import org.slf4j.LoggerFactory;
TTupleProtocol oprot = (TTupleProtocol) prot;
{
oprot.writeI32(struct.txn_ids.size());
- for (long _iter590 : struct.txn_ids)
+ for (long _iter598 : struct.txn_ids)
{
- oprot.writeI64(_iter590);
+ oprot.writeI64(_iter598);
}
}
}
@@ -421,13 +421,13 @@ import org.slf4j.LoggerFactory;
public void read(org.apache.thrift.protocol.TProtocol prot, AbortTxnsRequest struct) throws org.apache.thrift.TException {
TTupleProtocol iprot = (TTupleProtocol) prot;
{
- org.apache.thrift.protocol.TList _list591 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, iprot.readI32());
- struct.txn_ids = new ArrayList<Long>(_list591.size);
- long _elem592;
- for (int _i593 = 0; _i593 < _list591.size; ++_i593)
+ org.apache.thrift.protocol.TList _list599 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, iprot.readI32());
+ struct.txn_ids = new ArrayList<Long>(_list599.size);
+ long _elem600;
+ for (int _i601 = 0; _i601 < _list599.size; ++_i601)
{
- _elem592 = iprot.readI64();
- struct.txn_ids.add(_elem592);
+ _elem600 = iprot.readI64();
+ struct.txn_ids.add(_elem600);
}
}
struct.setTxn_idsIsSet(true);
http://git-wip-us.apache.org/repos/asf/hive/blob/1c9947f3/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddCheckConstraintRequest.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddCheckConstraintRequest.java b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddCheckConstraintRequest.java
index 02d552d..8ece410 100644
--- a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddCheckConstraintRequest.java
+++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddCheckConstraintRequest.java
@@ -354,14 +354,14 @@ import org.slf4j.LoggerFactory;
case 1: // CHECK_CONSTRAINT_COLS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list408 = iprot.readListBegin();
- struct.checkConstraintCols = new ArrayList<SQLCheckConstraint>(_list408.size);
- SQLCheckConstraint _elem409;
- for (int _i410 = 0; _i410 < _list408.size; ++_i410)
+ org.apache.thrift.protocol.TList _list416 = iprot.readListBegin();
+ struct.checkConstraintCols = new ArrayList<SQLCheckConstraint>(_list416.size);
+ SQLCheckConstraint _elem417;
+ for (int _i418 = 0; _i418 < _list416.size; ++_i418)
{
- _elem409 = new SQLCheckConstraint();
- _elem409.read(iprot);
- struct.checkConstraintCols.add(_elem409);
+ _elem417 = new SQLCheckConstraint();
+ _elem417.read(iprot);
+ struct.checkConstraintCols.add(_elem417);
}
iprot.readListEnd();
}
@@ -387,9 +387,9 @@ import org.slf4j.LoggerFactory;
oprot.writeFieldBegin(CHECK_CONSTRAINT_COLS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.checkConstraintCols.size()));
- for (SQLCheckConstraint _iter411 : struct.checkConstraintCols)
+ for (SQLCheckConstraint _iter419 : struct.checkConstraintCols)
{
- _iter411.write(oprot);
+ _iter419.write(oprot);
}
oprot.writeListEnd();
}
@@ -414,9 +414,9 @@ import org.slf4j.LoggerFactory;
TTupleProtocol oprot = (TTupleProtocol) prot;
{
oprot.writeI32(struct.checkConstraintCols.size());
- for (SQLCheckConstraint _iter412 : struct.checkConstraintCols)
+ for (SQLCheckConstraint _iter420 : struct.checkConstraintCols)
{
- _iter412.write(oprot);
+ _iter420.write(oprot);
}
}
}
@@ -425,14 +425,14 @@ import org.slf4j.LoggerFactory;
public void read(org.apache.thrift.protocol.TProtocol prot, AddCheckConstraintRequest struct) throws org.apache.thrift.TException {
TTupleProtocol iprot = (TTupleProtocol) prot;
{
- org.apache.thrift.protocol.TList _list413 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
- struct.checkConstraintCols = new ArrayList<SQLCheckConstraint>(_list413.size);
- SQLCheckConstraint _elem414;
- for (int _i415 = 0; _i415 < _list413.size; ++_i415)
+ org.apache.thrift.protocol.TList _list421 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
+ struct.checkConstraintCols = new ArrayList<SQLCheckConstraint>(_list421.size);
+ SQLCheckConstraint _elem422;
+ for (int _i423 = 0; _i423 < _list421.size; ++_i423)
{
- _elem414 = new SQLCheckConstraint();
- _elem414.read(iprot);
- struct.checkConstraintCols.add(_elem414);
+ _elem422 = new SQLCheckConstraint();
+ _elem422.read(iprot);
+ struct.checkConstraintCols.add(_elem422);
}
}
struct.setCheckConstraintColsIsSet(true);
http://git-wip-us.apache.org/repos/asf/hive/blob/1c9947f3/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddDefaultConstraintRequest.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddDefaultConstraintRequest.java b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddDefaultConstraintRequest.java
index 6acc6f8..8a125d8 100644
--- a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddDefaultConstraintRequest.java
+++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddDefaultConstraintRequest.java
@@ -354,14 +354,14 @@ import org.slf4j.LoggerFactory;
case 1: // DEFAULT_CONSTRAINT_COLS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list400 = iprot.readListBegin();
- struct.defaultConstraintCols = new ArrayList<SQLDefaultConstraint>(_list400.size);
- SQLDefaultConstraint _elem401;
- for (int _i402 = 0; _i402 < _list400.size; ++_i402)
+ org.apache.thrift.protocol.TList _list408 = iprot.readListBegin();
+ struct.defaultConstraintCols = new ArrayList<SQLDefaultConstraint>(_list408.size);
+ SQLDefaultConstraint _elem409;
+ for (int _i410 = 0; _i410 < _list408.size; ++_i410)
{
- _elem401 = new SQLDefaultConstraint();
- _elem401.read(iprot);
- struct.defaultConstraintCols.add(_elem401);
+ _elem409 = new SQLDefaultConstraint();
+ _elem409.read(iprot);
+ struct.defaultConstraintCols.add(_elem409);
}
iprot.readListEnd();
}
@@ -387,9 +387,9 @@ import org.slf4j.LoggerFactory;
oprot.writeFieldBegin(DEFAULT_CONSTRAINT_COLS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.defaultConstraintCols.size()));
- for (SQLDefaultConstraint _iter403 : struct.defaultConstraintCols)
+ for (SQLDefaultConstraint _iter411 : struct.defaultConstraintCols)
{
- _iter403.write(oprot);
+ _iter411.write(oprot);
}
oprot.writeListEnd();
}
@@ -414,9 +414,9 @@ import org.slf4j.LoggerFactory;
TTupleProtocol oprot = (TTupleProtocol) prot;
{
oprot.writeI32(struct.defaultConstraintCols.size());
- for (SQLDefaultConstraint _iter404 : struct.defaultConstraintCols)
+ for (SQLDefaultConstraint _iter412 : struct.defaultConstraintCols)
{
- _iter404.write(oprot);
+ _iter412.write(oprot);
}
}
}
@@ -425,14 +425,14 @@ import org.slf4j.LoggerFactory;
public void read(org.apache.thrift.protocol.TProtocol prot, AddDefaultConstraintRequest struct) throws org.apache.thrift.TException {
TTupleProtocol iprot = (TTupleProtocol) prot;
{
- org.apache.thrift.protocol.TList _list405 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
- struct.defaultConstraintCols = new ArrayList<SQLDefaultConstraint>(_list405.size);
- SQLDefaultConstraint _elem406;
- for (int _i407 = 0; _i407 < _list405.size; ++_i407)
+ org.apache.thrift.protocol.TList _list413 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
+ struct.defaultConstraintCols = new ArrayList<SQLDefaultConstraint>(_list413.size);
+ SQLDefaultConstraint _elem414;
+ for (int _i415 = 0; _i415 < _list413.size; ++_i415)
{
- _elem406 = new SQLDefaultConstraint();
- _elem406.read(iprot);
- struct.defaultConstraintCols.add(_elem406);
+ _elem414 = new SQLDefaultConstraint();
+ _elem414.read(iprot);
+ struct.defaultConstraintCols.add(_elem414);
}
}
struct.setDefaultConstraintColsIsSet(true);
http://git-wip-us.apache.org/repos/asf/hive/blob/1c9947f3/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddDynamicPartitions.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddDynamicPartitions.java b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddDynamicPartitions.java
index 3ce72e9..45618e7 100644
--- a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddDynamicPartitions.java
+++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddDynamicPartitions.java
@@ -816,13 +816,13 @@ import org.slf4j.LoggerFactory;
case 5: // PARTITIONNAMES
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list708 = iprot.readListBegin();
- struct.partitionnames = new ArrayList<String>(_list708.size);
- String _elem709;
- for (int _i710 = 0; _i710 < _list708.size; ++_i710)
+ org.apache.thrift.protocol.TList _list716 = iprot.readListBegin();
+ struct.partitionnames = new ArrayList<String>(_list716.size);
+ String _elem717;
+ for (int _i718 = 0; _i718 < _list716.size; ++_i718)
{
- _elem709 = iprot.readString();
- struct.partitionnames.add(_elem709);
+ _elem717 = iprot.readString();
+ struct.partitionnames.add(_elem717);
}
iprot.readListEnd();
}
@@ -872,9 +872,9 @@ import org.slf4j.LoggerFactory;
oprot.writeFieldBegin(PARTITIONNAMES_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.partitionnames.size()));
- for (String _iter711 : struct.partitionnames)
+ for (String _iter719 : struct.partitionnames)
{
- oprot.writeString(_iter711);
+ oprot.writeString(_iter719);
}
oprot.writeListEnd();
}
@@ -910,9 +910,9 @@ import org.slf4j.LoggerFactory;
oprot.writeString(struct.tablename);
{
oprot.writeI32(struct.partitionnames.size());
- for (String _iter712 : struct.partitionnames)
+ for (String _iter720 : struct.partitionnames)
{
- oprot.writeString(_iter712);
+ oprot.writeString(_iter720);
}
}
BitSet optionals = new BitSet();
@@ -937,13 +937,13 @@ import org.slf4j.LoggerFactory;
struct.tablename = iprot.readString();
struct.setTablenameIsSet(true);
{
- org.apache.thrift.protocol.TList _list713 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
- struct.partitionnames = new ArrayList<String>(_list713.size);
- String _elem714;
- for (int _i715 = 0; _i715 < _list713.size; ++_i715)
+ org.apache.thrift.protocol.TList _list721 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
+ struct.partitionnames = new ArrayList<String>(_list721.size);
+ String _elem722;
+ for (int _i723 = 0; _i723 < _list721.size; ++_i723)
{
- _elem714 = iprot.readString();
- struct.partitionnames.add(_elem714);
+ _elem722 = iprot.readString();
+ struct.partitionnames.add(_elem722);
}
}
struct.setPartitionnamesIsSet(true);
http://git-wip-us.apache.org/repos/asf/hive/blob/1c9947f3/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddForeignKeyRequest.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddForeignKeyRequest.java b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddForeignKeyRequest.java
index f57eb3b..0f443d4 100644
--- a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddForeignKeyRequest.java
+++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddForeignKeyRequest.java
@@ -354,14 +354,14 @@ import org.slf4j.LoggerFactory;
case 1: // FOREIGN_KEY_COLS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list376 = iprot.readListBegin();
- struct.foreignKeyCols = new ArrayList<SQLForeignKey>(_list376.size);
- SQLForeignKey _elem377;
- for (int _i378 = 0; _i378 < _list376.size; ++_i378)
+ org.apache.thrift.protocol.TList _list384 = iprot.readListBegin();
+ struct.foreignKeyCols = new ArrayList<SQLForeignKey>(_list384.size);
+ SQLForeignKey _elem385;
+ for (int _i386 = 0; _i386 < _list384.size; ++_i386)
{
- _elem377 = new SQLForeignKey();
- _elem377.read(iprot);
- struct.foreignKeyCols.add(_elem377);
+ _elem385 = new SQLForeignKey();
+ _elem385.read(iprot);
+ struct.foreignKeyCols.add(_elem385);
}
iprot.readListEnd();
}
@@ -387,9 +387,9 @@ import org.slf4j.LoggerFactory;
oprot.writeFieldBegin(FOREIGN_KEY_COLS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.foreignKeyCols.size()));
- for (SQLForeignKey _iter379 : struct.foreignKeyCols)
+ for (SQLForeignKey _iter387 : struct.foreignKeyCols)
{
- _iter379.write(oprot);
+ _iter387.write(oprot);
}
oprot.writeListEnd();
}
@@ -414,9 +414,9 @@ import org.slf4j.LoggerFactory;
TTupleProtocol oprot = (TTupleProtocol) prot;
{
oprot.writeI32(struct.foreignKeyCols.size());
- for (SQLForeignKey _iter380 : struct.foreignKeyCols)
+ for (SQLForeignKey _iter388 : struct.foreignKeyCols)
{
- _iter380.write(oprot);
+ _iter388.write(oprot);
}
}
}
@@ -425,14 +425,14 @@ import org.slf4j.LoggerFactory;
public void read(org.apache.thrift.protocol.TProtocol prot, AddForeignKeyRequest struct) throws org.apache.thrift.TException {
TTupleProtocol iprot = (TTupleProtocol) prot;
{
- org.apache.thrift.protocol.TList _list381 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
- struct.foreignKeyCols = new ArrayList<SQLForeignKey>(_list381.size);
- SQLForeignKey _elem382;
- for (int _i383 = 0; _i383 < _list381.size; ++_i383)
+ org.apache.thrift.protocol.TList _list389 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
+ struct.foreignKeyCols = new ArrayList<SQLForeignKey>(_list389.size);
+ SQLForeignKey _elem390;
+ for (int _i391 = 0; _i391 < _list389.size; ++_i391)
{
- _elem382 = new SQLForeignKey();
- _elem382.read(iprot);
- struct.foreignKeyCols.add(_elem382);
+ _elem390 = new SQLForeignKey();
+ _elem390.read(iprot);
+ struct.foreignKeyCols.add(_elem390);
}
}
struct.setForeignKeyColsIsSet(true);
http://git-wip-us.apache.org/repos/asf/hive/blob/1c9947f3/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddNotNullConstraintRequest.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddNotNullConstraintRequest.java b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddNotNullConstraintRequest.java
index e6bac16..0266bba 100644
--- a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddNotNullConstraintRequest.java
+++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddNotNullConstraintRequest.java
@@ -354,14 +354,14 @@ import org.slf4j.LoggerFactory;
case 1: // NOT_NULL_CONSTRAINT_COLS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list392 = iprot.readListBegin();
- struct.notNullConstraintCols = new ArrayList<SQLNotNullConstraint>(_list392.size);
- SQLNotNullConstraint _elem393;
- for (int _i394 = 0; _i394 < _list392.size; ++_i394)
+ org.apache.thrift.protocol.TList _list400 = iprot.readListBegin();
+ struct.notNullConstraintCols = new ArrayList<SQLNotNullConstraint>(_list400.size);
+ SQLNotNullConstraint _elem401;
+ for (int _i402 = 0; _i402 < _list400.size; ++_i402)
{
- _elem393 = new SQLNotNullConstraint();
- _elem393.read(iprot);
- struct.notNullConstraintCols.add(_elem393);
+ _elem401 = new SQLNotNullConstraint();
+ _elem401.read(iprot);
+ struct.notNullConstraintCols.add(_elem401);
}
iprot.readListEnd();
}
@@ -387,9 +387,9 @@ import org.slf4j.LoggerFactory;
oprot.writeFieldBegin(NOT_NULL_CONSTRAINT_COLS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.notNullConstraintCols.size()));
- for (SQLNotNullConstraint _iter395 : struct.notNullConstraintCols)
+ for (SQLNotNullConstraint _iter403 : struct.notNullConstraintCols)
{
- _iter395.write(oprot);
+ _iter403.write(oprot);
}
oprot.writeListEnd();
}
@@ -414,9 +414,9 @@ import org.slf4j.LoggerFactory;
TTupleProtocol oprot = (TTupleProtocol) prot;
{
oprot.writeI32(struct.notNullConstraintCols.size());
- for (SQLNotNullConstraint _iter396 : struct.notNullConstraintCols)
+ for (SQLNotNullConstraint _iter404 : struct.notNullConstraintCols)
{
- _iter396.write(oprot);
+ _iter404.write(oprot);
}
}
}
@@ -425,14 +425,14 @@ import org.slf4j.LoggerFactory;
public void read(org.apache.thrift.protocol.TProtocol prot, AddNotNullConstraintRequest struct) throws org.apache.thrift.TException {
TTupleProtocol iprot = (TTupleProtocol) prot;
{
- org.apache.thrift.protocol.TList _list397 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
- struct.notNullConstraintCols = new ArrayList<SQLNotNullConstraint>(_list397.size);
- SQLNotNullConstraint _elem398;
- for (int _i399 = 0; _i399 < _list397.size; ++_i399)
+ org.apache.thrift.protocol.TList _list405 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
+ struct.notNullConstraintCols = new ArrayList<SQLNotNullConstraint>(_list405.size);
+ SQLNotNullConstraint _elem406;
+ for (int _i407 = 0; _i407 < _list405.size; ++_i407)
{
- _elem398 = new SQLNotNullConstraint();
- _elem398.read(iprot);
- struct.notNullConstraintCols.add(_elem398);
+ _elem406 = new SQLNotNullConstraint();
+ _elem406.read(iprot);
+ struct.notNullConstraintCols.add(_elem406);
}
}
struct.setNotNullConstraintColsIsSet(true);
http://git-wip-us.apache.org/repos/asf/hive/blob/1c9947f3/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddPartitionsRequest.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddPartitionsRequest.java b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddPartitionsRequest.java
index 56e5043..ec42631 100644
--- a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddPartitionsRequest.java
+++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddPartitionsRequest.java
@@ -944,14 +944,14 @@ import org.slf4j.LoggerFactory;
case 3: // PARTS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list482 = iprot.readListBegin();
- struct.parts = new ArrayList<Partition>(_list482.size);
- Partition _elem483;
- for (int _i484 = 0; _i484 < _list482.size; ++_i484)
+ org.apache.thrift.protocol.TList _list490 = iprot.readListBegin();
+ struct.parts = new ArrayList<Partition>(_list490.size);
+ Partition _elem491;
+ for (int _i492 = 0; _i492 < _list490.size; ++_i492)
{
- _elem483 = new Partition();
- _elem483.read(iprot);
- struct.parts.add(_elem483);
+ _elem491 = new Partition();
+ _elem491.read(iprot);
+ struct.parts.add(_elem491);
}
iprot.readListEnd();
}
@@ -1027,9 +1027,9 @@ import org.slf4j.LoggerFactory;
oprot.writeFieldBegin(PARTS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.parts.size()));
- for (Partition _iter485 : struct.parts)
+ for (Partition _iter493 : struct.parts)
{
- _iter485.write(oprot);
+ _iter493.write(oprot);
}
oprot.writeListEnd();
}
@@ -1083,9 +1083,9 @@ import org.slf4j.LoggerFactory;
oprot.writeString(struct.tblName);
{
oprot.writeI32(struct.parts.size());
- for (Partition _iter486 : struct.parts)
+ for (Partition _iter494 : struct.parts)
{
- _iter486.write(oprot);
+ _iter494.write(oprot);
}
}
oprot.writeBool(struct.ifNotExists);
@@ -1125,14 +1125,14 @@ import org.slf4j.LoggerFactory;
struct.tblName = iprot.readString();
struct.setTblNameIsSet(true);
{
- org.apache.thrift.protocol.TList _list487 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
- struct.parts = new ArrayList<Partition>(_list487.size);
- Partition _elem488;
- for (int _i489 = 0; _i489 < _list487.size; ++_i489)
+ org.apache.thrift.protocol.TList _list495 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
+ struct.parts = new ArrayList<Partition>(_list495.size);
+ Partition _elem496;
+ for (int _i497 = 0; _i497 < _list495.size; ++_i497)
{
- _elem488 = new Partition();
- _elem488.read(iprot);
- struct.parts.add(_elem488);
+ _elem496 = new Partition();
+ _elem496.read(iprot);
+ struct.parts.add(_elem496);
}
}
struct.setPartsIsSet(true);
http://git-wip-us.apache.org/repos/asf/hive/blob/1c9947f3/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddPartitionsResult.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddPartitionsResult.java b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddPartitionsResult.java
index 3c0bf82..a00af60 100644
--- a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddPartitionsResult.java
+++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddPartitionsResult.java
@@ -426,14 +426,14 @@ import org.slf4j.LoggerFactory;
case 1: // PARTITIONS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list474 = iprot.readListBegin();
- struct.partitions = new ArrayList<Partition>(_list474.size);
- Partition _elem475;
- for (int _i476 = 0; _i476 < _list474.size; ++_i476)
+ org.apache.thrift.protocol.TList _list482 = iprot.readListBegin();
+ struct.partitions = new ArrayList<Partition>(_list482.size);
+ Partition _elem483;
+ for (int _i484 = 0; _i484 < _list482.size; ++_i484)
{
- _elem475 = new Partition();
- _elem475.read(iprot);
- struct.partitions.add(_elem475);
+ _elem483 = new Partition();
+ _elem483.read(iprot);
+ struct.partitions.add(_elem483);
}
iprot.readListEnd();
}
@@ -468,9 +468,9 @@ import org.slf4j.LoggerFactory;
oprot.writeFieldBegin(PARTITIONS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.partitions.size()));
- for (Partition _iter477 : struct.partitions)
+ for (Partition _iter485 : struct.partitions)
{
- _iter477.write(oprot);
+ _iter485.write(oprot);
}
oprot.writeListEnd();
}
@@ -510,9 +510,9 @@ import org.slf4j.LoggerFactory;
if (struct.isSetPartitions()) {
{
oprot.writeI32(struct.partitions.size());
- for (Partition _iter478 : struct.partitions)
+ for (Partition _iter486 : struct.partitions)
{
- _iter478.write(oprot);
+ _iter486.write(oprot);
}
}
}
@@ -527,14 +527,14 @@ import org.slf4j.LoggerFactory;
BitSet incoming = iprot.readBitSet(2);
if (incoming.get(0)) {
{
- org.apache.thrift.protocol.TList _list479 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
- struct.partitions = new ArrayList<Partition>(_list479.size);
- Partition _elem480;
- for (int _i481 = 0; _i481 < _list479.size; ++_i481)
+ org.apache.thrift.protocol.TList _list487 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
+ struct.partitions = new ArrayList<Partition>(_list487.size);
+ Partition _elem488;
+ for (int _i489 = 0; _i489 < _list487.size; ++_i489)
{
- _elem480 = new Partition();
- _elem480.read(iprot);
- struct.partitions.add(_elem480);
+ _elem488 = new Partition();
+ _elem488.read(iprot);
+ struct.partitions.add(_elem488);
}
}
struct.setPartitionsIsSet(true);
http://git-wip-us.apache.org/repos/asf/hive/blob/1c9947f3/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddPrimaryKeyRequest.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddPrimaryKeyRequest.java b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddPrimaryKeyRequest.java
index 39bb6be..9069a41 100644
--- a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddPrimaryKeyRequest.java
+++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddPrimaryKeyRequest.java
@@ -354,14 +354,14 @@ import org.slf4j.LoggerFactory;
case 1: // PRIMARY_KEY_COLS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list368 = iprot.readListBegin();
- struct.primaryKeyCols = new ArrayList<SQLPrimaryKey>(_list368.size);
- SQLPrimaryKey _elem369;
- for (int _i370 = 0; _i370 < _list368.size; ++_i370)
+ org.apache.thrift.protocol.TList _list376 = iprot.readListBegin();
+ struct.primaryKeyCols = new ArrayList<SQLPrimaryKey>(_list376.size);
+ SQLPrimaryKey _elem377;
+ for (int _i378 = 0; _i378 < _list376.size; ++_i378)
{
- _elem369 = new SQLPrimaryKey();
- _elem369.read(iprot);
- struct.primaryKeyCols.add(_elem369);
+ _elem377 = new SQLPrimaryKey();
+ _elem377.read(iprot);
+ struct.primaryKeyCols.add(_elem377);
}
iprot.readListEnd();
}
@@ -387,9 +387,9 @@ import org.slf4j.LoggerFactory;
oprot.writeFieldBegin(PRIMARY_KEY_COLS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.primaryKeyCols.size()));
- for (SQLPrimaryKey _iter371 : struct.primaryKeyCols)
+ for (SQLPrimaryKey _iter379 : struct.primaryKeyCols)
{
- _iter371.write(oprot);
+ _iter379.write(oprot);
}
oprot.writeListEnd();
}
@@ -414,9 +414,9 @@ import org.slf4j.LoggerFactory;
TTupleProtocol oprot = (TTupleProtocol) prot;
{
oprot.writeI32(struct.primaryKeyCols.size());
- for (SQLPrimaryKey _iter372 : struct.primaryKeyCols)
+ for (SQLPrimaryKey _iter380 : struct.primaryKeyCols)
{
- _iter372.write(oprot);
+ _iter380.write(oprot);
}
}
}
@@ -425,14 +425,14 @@ import org.slf4j.LoggerFactory;
public void read(org.apache.thrift.protocol.TProtocol prot, AddPrimaryKeyRequest struct) throws org.apache.thrift.TException {
TTupleProtocol iprot = (TTupleProtocol) prot;
{
- org.apache.thrift.protocol.TList _list373 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
- struct.primaryKeyCols = new ArrayList<SQLPrimaryKey>(_list373.size);
- SQLPrimaryKey _elem374;
- for (int _i375 = 0; _i375 < _list373.size; ++_i375)
+ org.apache.thrift.protocol.TList _list381 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
+ struct.primaryKeyCols = new ArrayList<SQLPrimaryKey>(_list381.size);
+ SQLPrimaryKey _elem382;
+ for (int _i383 = 0; _i383 < _list381.size; ++_i383)
{
- _elem374 = new SQLPrimaryKey();
- _elem374.read(iprot);
- struct.primaryKeyCols.add(_elem374);
+ _elem382 = new SQLPrimaryKey();
+ _elem382.read(iprot);
+ struct.primaryKeyCols.add(_elem382);
}
}
struct.setPrimaryKeyColsIsSet(true);
http://git-wip-us.apache.org/repos/asf/hive/blob/1c9947f3/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddUniqueConstraintRequest.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddUniqueConstraintRequest.java b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddUniqueConstraintRequest.java
index bcb1e6b..c47db4a 100644
--- a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddUniqueConstraintRequest.java
+++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddUniqueConstraintRequest.java
@@ -354,14 +354,14 @@ import org.slf4j.LoggerFactory;
case 1: // UNIQUE_CONSTRAINT_COLS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list384 = iprot.readListBegin();
- struct.uniqueConstraintCols = new ArrayList<SQLUniqueConstraint>(_list384.size);
- SQLUniqueConstraint _elem385;
- for (int _i386 = 0; _i386 < _list384.size; ++_i386)
+ org.apache.thrift.protocol.TList _list392 = iprot.readListBegin();
+ struct.uniqueConstraintCols = new ArrayList<SQLUniqueConstraint>(_list392.size);
+ SQLUniqueConstraint _elem393;
+ for (int _i394 = 0; _i394 < _list392.size; ++_i394)
{
- _elem385 = new SQLUniqueConstraint();
- _elem385.read(iprot);
- struct.uniqueConstraintCols.add(_elem385);
+ _elem393 = new SQLUniqueConstraint();
+ _elem393.read(iprot);
+ struct.uniqueConstraintCols.add(_elem393);
}
iprot.readListEnd();
}
@@ -387,9 +387,9 @@ import org.slf4j.LoggerFactory;
oprot.writeFieldBegin(UNIQUE_CONSTRAINT_COLS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.uniqueConstraintCols.size()));
- for (SQLUniqueConstraint _iter387 : struct.uniqueConstraintCols)
+ for (SQLUniqueConstraint _iter395 : struct.uniqueConstraintCols)
{
- _iter387.write(oprot);
+ _iter395.write(oprot);
}
oprot.writeListEnd();
}
@@ -414,9 +414,9 @@ import org.slf4j.LoggerFactory;
TTupleProtocol oprot = (TTupleProtocol) prot;
{
oprot.writeI32(struct.uniqueConstraintCols.size());
- for (SQLUniqueConstraint _iter388 : struct.uniqueConstraintCols)
+ for (SQLUniqueConstraint _iter396 : struct.uniqueConstraintCols)
{
- _iter388.write(oprot);
+ _iter396.write(oprot);
}
}
}
@@ -425,14 +425,14 @@ import org.slf4j.LoggerFactory;
public void read(org.apache.thrift.protocol.TProtocol prot, AddUniqueConstraintRequest struct) throws org.apache.thrift.TException {
TTupleProtocol iprot = (TTupleProtocol) prot;
{
- org.apache.thrift.protocol.TList _list389 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
- struct.uniqueConstraintCols = new ArrayList<SQLUniqueConstraint>(_list389.size);
- SQLUniqueConstraint _elem390;
- for (int _i391 = 0; _i391 < _list389.size; ++_i391)
+ org.apache.thrift.protocol.TList _list397 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
+ struct.uniqueConstraintCols = new ArrayList<SQLUniqueConstraint>(_list397.size);
+ SQLUniqueConstraint _elem398;
+ for (int _i399 = 0; _i399 < _list397.size; ++_i399)
{
- _elem390 = new SQLUniqueConstraint();
- _elem390.read(iprot);
- struct.uniqueConstraintCols.add(_elem390);
+ _elem398 = new SQLUniqueConstraint();
+ _elem398.read(iprot);
+ struct.uniqueConstraintCols.add(_elem398);
}
}
struct.setUniqueConstraintColsIsSet(true);
http://git-wip-us.apache.org/repos/asf/hive/blob/1c9947f3/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AggrStats.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AggrStats.java b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AggrStats.java
index 87b8fea..67d9b10 100644
--- a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AggrStats.java
+++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AggrStats.java
@@ -516,14 +516,14 @@ import org.slf4j.LoggerFactory;
case 1: // COL_STATS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list276 = iprot.readListBegin();
- struct.colStats = new ArrayList<ColumnStatisticsObj>(_list276.size);
- ColumnStatisticsObj _elem277;
- for (int _i278 = 0; _i278 < _list276.size; ++_i278)
+ org.apache.thrift.protocol.TList _list284 = iprot.readListBegin();
+ struct.colStats = new ArrayList<ColumnStatisticsObj>(_list284.size);
+ ColumnStatisticsObj _elem285;
+ for (int _i286 = 0; _i286 < _list284.size; ++_i286)
{
- _elem277 = new ColumnStatisticsObj();
- _elem277.read(iprot);
- struct.colStats.add(_elem277);
+ _elem285 = new ColumnStatisticsObj();
+ _elem285.read(iprot);
+ struct.colStats.add(_elem285);
}
iprot.readListEnd();
}
@@ -565,9 +565,9 @@ import org.slf4j.LoggerFactory;
oprot.writeFieldBegin(COL_STATS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.colStats.size()));
- for (ColumnStatisticsObj _iter279 : struct.colStats)
+ for (ColumnStatisticsObj _iter287 : struct.colStats)
{
- _iter279.write(oprot);
+ _iter287.write(oprot);
}
oprot.writeListEnd();
}
@@ -600,9 +600,9 @@ import org.slf4j.LoggerFactory;
TTupleProtocol oprot = (TTupleProtocol) prot;
{
oprot.writeI32(struct.colStats.size());
- for (ColumnStatisticsObj _iter280 : struct.colStats)
+ for (ColumnStatisticsObj _iter288 : struct.colStats)
{
- _iter280.write(oprot);
+ _iter288.write(oprot);
}
}
oprot.writeI64(struct.partsFound);
@@ -620,14 +620,14 @@ import org.slf4j.LoggerFactory;
public void read(org.apache.thrift.protocol.TProtocol prot, AggrStats struct) throws org.apache.thrift.TException {
TTupleProtocol iprot = (TTupleProtocol) prot;
{
- org.apache.thrift.protocol.TList _list281 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
- struct.colStats = new ArrayList<ColumnStatisticsObj>(_list281.size);
- ColumnStatisticsObj _elem282;
- for (int _i283 = 0; _i283 < _list281.size; ++_i283)
+ org.apache.thrift.protocol.TList _list289 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
+ struct.colStats = new ArrayList<ColumnStatisticsObj>(_list289.size);
+ ColumnStatisticsObj _elem290;
+ for (int _i291 = 0; _i291 < _list289.size; ++_i291)
{
- _elem282 = new ColumnStatisticsObj();
- _elem282.read(iprot);
- struct.colStats.add(_elem282);
+ _elem290 = new ColumnStatisticsObj();
+ _elem290.read(iprot);
+ struct.colStats.add(_elem290);
}
}
struct.setColStatsIsSet(true);
http://git-wip-us.apache.org/repos/asf/hive/blob/1c9947f3/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AllocateTableWriteIdsRequest.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AllocateTableWriteIdsRequest.java b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AllocateTableWriteIdsRequest.java
index a0b47a9..5fcb98f 100644
--- a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AllocateTableWriteIdsRequest.java
+++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AllocateTableWriteIdsRequest.java
@@ -716,13 +716,13 @@ import org.slf4j.LoggerFactory;
case 3: // TXN_IDS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list634 = iprot.readListBegin();
- struct.txnIds = new ArrayList<Long>(_list634.size);
- long _elem635;
- for (int _i636 = 0; _i636 < _list634.size; ++_i636)
+ org.apache.thrift.protocol.TList _list642 = iprot.readListBegin();
+ struct.txnIds = new ArrayList<Long>(_list642.size);
+ long _elem643;
+ for (int _i644 = 0; _i644 < _list642.size; ++_i644)
{
- _elem635 = iprot.readI64();
- struct.txnIds.add(_elem635);
+ _elem643 = iprot.readI64();
+ struct.txnIds.add(_elem643);
}
iprot.readListEnd();
}
@@ -742,14 +742,14 @@ import org.slf4j.LoggerFactory;
case 5: // SRC_TXN_TO_WRITE_ID_LIST
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list637 = iprot.readListBegin();
- struct.srcTxnToWriteIdList = new ArrayList<TxnToWriteId>(_list637.size);
- TxnToWriteId _elem638;
- for (int _i639 = 0; _i639 < _list637.size; ++_i639)
+ org.apache.thrift.protocol.TList _list645 = iprot.readListBegin();
+ struct.srcTxnToWriteIdList = new ArrayList<TxnToWriteId>(_list645.size);
+ TxnToWriteId _elem646;
+ for (int _i647 = 0; _i647 < _list645.size; ++_i647)
{
- _elem638 = new TxnToWriteId();
- _elem638.read(iprot);
- struct.srcTxnToWriteIdList.add(_elem638);
+ _elem646 = new TxnToWriteId();
+ _elem646.read(iprot);
+ struct.srcTxnToWriteIdList.add(_elem646);
}
iprot.readListEnd();
}
@@ -786,9 +786,9 @@ import org.slf4j.LoggerFactory;
oprot.writeFieldBegin(TXN_IDS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, struct.txnIds.size()));
- for (long _iter640 : struct.txnIds)
+ for (long _iter648 : struct.txnIds)
{
- oprot.writeI64(_iter640);
+ oprot.writeI64(_iter648);
}
oprot.writeListEnd();
}
@@ -807,9 +807,9 @@ import org.slf4j.LoggerFactory;
oprot.writeFieldBegin(SRC_TXN_TO_WRITE_ID_LIST_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.srcTxnToWriteIdList.size()));
- for (TxnToWriteId _iter641 : struct.srcTxnToWriteIdList)
+ for (TxnToWriteId _iter649 : struct.srcTxnToWriteIdList)
{
- _iter641.write(oprot);
+ _iter649.write(oprot);
}
oprot.writeListEnd();
}
@@ -849,9 +849,9 @@ import org.slf4j.LoggerFactory;
if (struct.isSetTxnIds()) {
{
oprot.writeI32(struct.txnIds.size());
- for (long _iter642 : struct.txnIds)
+ for (long _iter650 : struct.txnIds)
{
- oprot.writeI64(_iter642);
+ oprot.writeI64(_iter650);
}
}
}
@@ -861,9 +861,9 @@ import org.slf4j.LoggerFactory;
if (struct.isSetSrcTxnToWriteIdList()) {
{
oprot.writeI32(struct.srcTxnToWriteIdList.size());
- for (TxnToWriteId _iter643 : struct.srcTxnToWriteIdList)
+ for (TxnToWriteId _iter651 : struct.srcTxnToWriteIdList)
{
- _iter643.write(oprot);
+ _iter651.write(oprot);
}
}
}
@@ -879,13 +879,13 @@ import org.slf4j.LoggerFactory;
BitSet incoming = iprot.readBitSet(3);
if (incoming.get(0)) {
{
- org.apache.thrift.protocol.TList _list644 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, iprot.readI32());
- struct.txnIds = new ArrayList<Long>(_list644.size);
- long _elem645;
- for (int _i646 = 0; _i646 < _list644.size; ++_i646)
+ org.apache.thrift.protocol.TList _list652 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, iprot.readI32());
+ struct.txnIds = new ArrayList<Long>(_list652.size);
+ long _elem653;
+ for (int _i654 = 0; _i654 < _list652.size; ++_i654)
{
- _elem645 = iprot.readI64();
- struct.txnIds.add(_elem645);
+ _elem653 = iprot.readI64();
+ struct.txnIds.add(_elem653);
}
}
struct.setTxnIdsIsSet(true);
@@ -896,14 +896,14 @@ import org.slf4j.LoggerFactory;
}
if (incoming.get(2)) {
{
- org.apache.thrift.protocol.TList _list647 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
- struct.srcTxnToWriteIdList = new ArrayList<TxnToWriteId>(_list647.size);
- TxnToWriteId _elem648;
- for (int _i649 = 0; _i649 < _list647.size; ++_i649)
+ org.apache.thrift.protocol.TList _list655 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
+ struct.srcTxnToWriteIdList = new ArrayList<TxnToWriteId>(_list655.size);
+ TxnToWriteId _elem656;
+ for (int _i657 = 0; _i657 < _list655.size; ++_i657)
{
- _elem648 = new TxnToWriteId();
- _elem648.read(iprot);
- struct.srcTxnToWriteIdList.add(_elem648);
+ _elem656 = new TxnToWriteId();
+ _elem656.read(iprot);
+ struct.srcTxnToWriteIdList.add(_elem656);
}
}
struct.setSrcTxnToWriteIdListIsSet(true);
http://git-wip-us.apache.org/repos/asf/hive/blob/1c9947f3/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AllocateTableWriteIdsResponse.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AllocateTableWriteIdsResponse.java b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AllocateTableWriteIdsResponse.java
index 13df26d..2a13eba 100644
--- a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AllocateTableWriteIdsResponse.java
+++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AllocateTableWriteIdsResponse.java
@@ -354,14 +354,14 @@ import org.slf4j.LoggerFactory;
case 1: // TXN_TO_WRITE_IDS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list650 = iprot.readListBegin();
- struct.txnToWriteIds = new ArrayList<TxnToWriteId>(_list650.size);
- TxnToWriteId _elem651;
- for (int _i652 = 0; _i652 < _list650.size; ++_i652)
+ org.apache.thrift.protocol.TList _list658 = iprot.readListBegin();
+ struct.txnToWriteIds = new ArrayList<TxnToWriteId>(_list658.size);
+ TxnToWriteId _elem659;
+ for (int _i660 = 0; _i660 < _list658.size; ++_i660)
{
- _elem651 = new TxnToWriteId();
- _elem651.read(iprot);
- struct.txnToWriteIds.add(_elem651);
+ _elem659 = new TxnToWriteId();
+ _elem659.read(iprot);
+ struct.txnToWriteIds.add(_elem659);
}
iprot.readListEnd();
}
@@ -387,9 +387,9 @@ import org.slf4j.LoggerFactory;
oprot.writeFieldBegin(TXN_TO_WRITE_IDS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.txnToWriteIds.size()));
- for (TxnToWriteId _iter653 : struct.txnToWriteIds)
+ for (TxnToWriteId _iter661 : struct.txnToWriteIds)
{
- _iter653.write(oprot);
+ _iter661.write(oprot);
}
oprot.writeListEnd();
}
@@ -414,9 +414,9 @@ import org.slf4j.LoggerFactory;
TTupleProtocol oprot = (TTupleProtocol) prot;
{
oprot.writeI32(struct.txnToWriteIds.size());
- for (TxnToWriteId _iter654 : struct.txnToWriteIds)
+ for (TxnToWriteId _iter662 : struct.txnToWriteIds)
{
- _iter654.write(oprot);
+ _iter662.write(oprot);
}
}
}
@@ -425,14 +425,14 @@ import org.slf4j.LoggerFactory;
public void read(org.apache.thrift.protocol.TProtocol prot, AllocateTableWriteIdsResponse struct) throws org.apache.thrift.TException {
TTupleProtocol iprot = (TTupleProtocol) prot;
{
- org.apache.thrift.protocol.TList _list655 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
- struct.txnToWriteIds = new ArrayList<TxnToWriteId>(_list655.size);
- TxnToWriteId _elem656;
- for (int _i657 = 0; _i657 < _list655.size; ++_i657)
+ org.apache.thrift.protocol.TList _list663 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
+ struct.txnToWriteIds = new ArrayList<TxnToWriteId>(_list663.size);
+ TxnToWriteId _elem664;
+ for (int _i665 = 0; _i665 < _list663.size; ++_i665)
{
- _elem656 = new TxnToWriteId();
- _elem656.read(iprot);
- struct.txnToWriteIds.add(_elem656);
+ _elem664 = new TxnToWriteId();
+ _elem664.read(iprot);
+ struct.txnToWriteIds.add(_elem664);
}
}
struct.setTxnToWriteIdsIsSet(true);