You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by se...@apache.org on 2018/07/25 18:27:41 UTC
[32/50] [abbrv] hive git commit: HIVE-20047 : remove txnID argument
for txn stats methods (Sergey Shelukhin)
http://git-wip-us.apache.org/repos/asf/hive/blob/cdb32a7f/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ttypes.py
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ttypes.py b/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ttypes.py
index 923fe2c..7fc1e43 100644
--- a/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ttypes.py
+++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ttypes.py
@@ -2510,7 +2510,6 @@ class TruncateTableRequest:
- dbName
- tableName
- partNames
- - txnId
- writeId
- validWriteIdList
"""
@@ -2520,16 +2519,14 @@ class TruncateTableRequest:
(1, TType.STRING, 'dbName', None, None, ), # 1
(2, TType.STRING, 'tableName', None, None, ), # 2
(3, TType.LIST, 'partNames', (TType.STRING,None), None, ), # 3
- (4, TType.I64, 'txnId', None, -1, ), # 4
- (5, TType.I64, 'writeId', None, -1, ), # 5
- (6, TType.STRING, 'validWriteIdList', None, None, ), # 6
+ (4, TType.I64, 'writeId', None, -1, ), # 4
+ (5, TType.STRING, 'validWriteIdList', None, None, ), # 5
)
- def __init__(self, dbName=None, tableName=None, partNames=None, txnId=thrift_spec[4][4], writeId=thrift_spec[5][4], validWriteIdList=None,):
+ def __init__(self, dbName=None, tableName=None, partNames=None, writeId=thrift_spec[4][4], validWriteIdList=None,):
self.dbName = dbName
self.tableName = tableName
self.partNames = partNames
- self.txnId = txnId
self.writeId = writeId
self.validWriteIdList = validWriteIdList
@@ -2564,15 +2561,10 @@ class TruncateTableRequest:
iprot.skip(ftype)
elif fid == 4:
if ftype == TType.I64:
- self.txnId = iprot.readI64()
- else:
- iprot.skip(ftype)
- elif fid == 5:
- if ftype == TType.I64:
self.writeId = iprot.readI64()
else:
iprot.skip(ftype)
- elif fid == 6:
+ elif fid == 5:
if ftype == TType.STRING:
self.validWriteIdList = iprot.readString()
else:
@@ -2602,16 +2594,12 @@ class TruncateTableRequest:
oprot.writeString(iter75)
oprot.writeListEnd()
oprot.writeFieldEnd()
- if self.txnId is not None:
- oprot.writeFieldBegin('txnId', TType.I64, 4)
- oprot.writeI64(self.txnId)
- oprot.writeFieldEnd()
if self.writeId is not None:
- oprot.writeFieldBegin('writeId', TType.I64, 5)
+ oprot.writeFieldBegin('writeId', TType.I64, 4)
oprot.writeI64(self.writeId)
oprot.writeFieldEnd()
if self.validWriteIdList is not None:
- oprot.writeFieldBegin('validWriteIdList', TType.STRING, 6)
+ oprot.writeFieldBegin('validWriteIdList', TType.STRING, 5)
oprot.writeString(self.validWriteIdList)
oprot.writeFieldEnd()
oprot.writeFieldStop()
@@ -2630,7 +2618,6 @@ class TruncateTableRequest:
value = (value * 31) ^ hash(self.dbName)
value = (value * 31) ^ hash(self.tableName)
value = (value * 31) ^ hash(self.partNames)
- value = (value * 31) ^ hash(self.txnId)
value = (value * 31) ^ hash(self.writeId)
value = (value * 31) ^ hash(self.validWriteIdList)
return value
@@ -7316,7 +7303,6 @@ class SetPartitionsStatsRequest:
Attributes:
- colStats
- needMerge
- - txnId
- writeId
- validWriteIdList
"""
@@ -7325,15 +7311,13 @@ class SetPartitionsStatsRequest:
None, # 0
(1, TType.LIST, 'colStats', (TType.STRUCT,(ColumnStatistics, ColumnStatistics.thrift_spec)), None, ), # 1
(2, TType.BOOL, 'needMerge', None, None, ), # 2
- (3, TType.I64, 'txnId', None, -1, ), # 3
- (4, TType.I64, 'writeId', None, -1, ), # 4
- (5, TType.STRING, 'validWriteIdList', None, None, ), # 5
+ (3, TType.I64, 'writeId', None, -1, ), # 3
+ (4, TType.STRING, 'validWriteIdList', None, None, ), # 4
)
- def __init__(self, colStats=None, needMerge=None, txnId=thrift_spec[3][4], writeId=thrift_spec[4][4], validWriteIdList=None,):
+ def __init__(self, colStats=None, needMerge=None, writeId=thrift_spec[3][4], validWriteIdList=None,):
self.colStats = colStats
self.needMerge = needMerge
- self.txnId = txnId
self.writeId = writeId
self.validWriteIdList = validWriteIdList
@@ -7364,15 +7348,10 @@ class SetPartitionsStatsRequest:
iprot.skip(ftype)
elif fid == 3:
if ftype == TType.I64:
- self.txnId = iprot.readI64()
- else:
- iprot.skip(ftype)
- elif fid == 4:
- if ftype == TType.I64:
self.writeId = iprot.readI64()
else:
iprot.skip(ftype)
- elif fid == 5:
+ elif fid == 4:
if ftype == TType.STRING:
self.validWriteIdList = iprot.readString()
else:
@@ -7398,16 +7377,12 @@ class SetPartitionsStatsRequest:
oprot.writeFieldBegin('needMerge', TType.BOOL, 2)
oprot.writeBool(self.needMerge)
oprot.writeFieldEnd()
- if self.txnId is not None:
- oprot.writeFieldBegin('txnId', TType.I64, 3)
- oprot.writeI64(self.txnId)
- oprot.writeFieldEnd()
if self.writeId is not None:
- oprot.writeFieldBegin('writeId', TType.I64, 4)
+ oprot.writeFieldBegin('writeId', TType.I64, 3)
oprot.writeI64(self.writeId)
oprot.writeFieldEnd()
if self.validWriteIdList is not None:
- oprot.writeFieldBegin('validWriteIdList', TType.STRING, 5)
+ oprot.writeFieldBegin('validWriteIdList', TType.STRING, 4)
oprot.writeString(self.validWriteIdList)
oprot.writeFieldEnd()
oprot.writeFieldStop()
@@ -7423,7 +7398,6 @@ class SetPartitionsStatsRequest:
value = 17
value = (value * 31) ^ hash(self.colStats)
value = (value * 31) ^ hash(self.needMerge)
- value = (value * 31) ^ hash(self.txnId)
value = (value * 31) ^ hash(self.writeId)
value = (value * 31) ^ hash(self.validWriteIdList)
return value
@@ -9722,7 +9696,6 @@ class TableStatsRequest:
- tblName
- colNames
- catName
- - txnId
- validWriteIdList
"""
@@ -9732,16 +9705,14 @@ class TableStatsRequest:
(2, TType.STRING, 'tblName', None, None, ), # 2
(3, TType.LIST, 'colNames', (TType.STRING,None), None, ), # 3
(4, TType.STRING, 'catName', None, None, ), # 4
- (5, TType.I64, 'txnId', None, -1, ), # 5
- (6, TType.STRING, 'validWriteIdList', None, None, ), # 6
+ (5, TType.STRING, 'validWriteIdList', None, None, ), # 5
)
- def __init__(self, dbName=None, tblName=None, colNames=None, catName=None, txnId=thrift_spec[5][4], validWriteIdList=None,):
+ def __init__(self, dbName=None, tblName=None, colNames=None, catName=None, validWriteIdList=None,):
self.dbName = dbName
self.tblName = tblName
self.colNames = colNames
self.catName = catName
- self.txnId = txnId
self.validWriteIdList = validWriteIdList
def read(self, iprot):
@@ -9779,11 +9750,6 @@ class TableStatsRequest:
else:
iprot.skip(ftype)
elif fid == 5:
- if ftype == TType.I64:
- self.txnId = iprot.readI64()
- else:
- iprot.skip(ftype)
- elif fid == 6:
if ftype == TType.STRING:
self.validWriteIdList = iprot.readString()
else:
@@ -9817,12 +9783,8 @@ class TableStatsRequest:
oprot.writeFieldBegin('catName', TType.STRING, 4)
oprot.writeString(self.catName)
oprot.writeFieldEnd()
- if self.txnId is not None:
- oprot.writeFieldBegin('txnId', TType.I64, 5)
- oprot.writeI64(self.txnId)
- oprot.writeFieldEnd()
if self.validWriteIdList is not None:
- oprot.writeFieldBegin('validWriteIdList', TType.STRING, 6)
+ oprot.writeFieldBegin('validWriteIdList', TType.STRING, 5)
oprot.writeString(self.validWriteIdList)
oprot.writeFieldEnd()
oprot.writeFieldStop()
@@ -9844,7 +9806,6 @@ class TableStatsRequest:
value = (value * 31) ^ hash(self.tblName)
value = (value * 31) ^ hash(self.colNames)
value = (value * 31) ^ hash(self.catName)
- value = (value * 31) ^ hash(self.txnId)
value = (value * 31) ^ hash(self.validWriteIdList)
return value
@@ -9867,7 +9828,6 @@ class PartitionsStatsRequest:
- colNames
- partNames
- catName
- - txnId
- validWriteIdList
"""
@@ -9878,17 +9838,15 @@ class PartitionsStatsRequest:
(3, TType.LIST, 'colNames', (TType.STRING,None), None, ), # 3
(4, TType.LIST, 'partNames', (TType.STRING,None), None, ), # 4
(5, TType.STRING, 'catName', None, None, ), # 5
- (6, TType.I64, 'txnId', None, -1, ), # 6
- (7, TType.STRING, 'validWriteIdList', None, None, ), # 7
+ (6, TType.STRING, 'validWriteIdList', None, None, ), # 6
)
- def __init__(self, dbName=None, tblName=None, colNames=None, partNames=None, catName=None, txnId=thrift_spec[6][4], validWriteIdList=None,):
+ def __init__(self, dbName=None, tblName=None, colNames=None, partNames=None, catName=None, validWriteIdList=None,):
self.dbName = dbName
self.tblName = tblName
self.colNames = colNames
self.partNames = partNames
self.catName = catName
- self.txnId = txnId
self.validWriteIdList = validWriteIdList
def read(self, iprot):
@@ -9936,11 +9894,6 @@ class PartitionsStatsRequest:
else:
iprot.skip(ftype)
elif fid == 6:
- if ftype == TType.I64:
- self.txnId = iprot.readI64()
- else:
- iprot.skip(ftype)
- elif fid == 7:
if ftype == TType.STRING:
self.validWriteIdList = iprot.readString()
else:
@@ -9981,12 +9934,8 @@ class PartitionsStatsRequest:
oprot.writeFieldBegin('catName', TType.STRING, 5)
oprot.writeString(self.catName)
oprot.writeFieldEnd()
- if self.txnId is not None:
- oprot.writeFieldBegin('txnId', TType.I64, 6)
- oprot.writeI64(self.txnId)
- oprot.writeFieldEnd()
if self.validWriteIdList is not None:
- oprot.writeFieldBegin('validWriteIdList', TType.STRING, 7)
+ oprot.writeFieldBegin('validWriteIdList', TType.STRING, 6)
oprot.writeString(self.validWriteIdList)
oprot.writeFieldEnd()
oprot.writeFieldStop()
@@ -10011,7 +9960,6 @@ class PartitionsStatsRequest:
value = (value * 31) ^ hash(self.colNames)
value = (value * 31) ^ hash(self.partNames)
value = (value * 31) ^ hash(self.catName)
- value = (value * 31) ^ hash(self.txnId)
value = (value * 31) ^ hash(self.validWriteIdList)
return value
@@ -10122,7 +10070,6 @@ class AddPartitionsRequest:
- ifNotExists
- needResult
- catName
- - txnId
- validWriteIdList
"""
@@ -10134,18 +10081,16 @@ class AddPartitionsRequest:
(4, TType.BOOL, 'ifNotExists', None, None, ), # 4
(5, TType.BOOL, 'needResult', None, True, ), # 5
(6, TType.STRING, 'catName', None, None, ), # 6
- (7, TType.I64, 'txnId', None, -1, ), # 7
- (8, TType.STRING, 'validWriteIdList', None, None, ), # 8
+ (7, TType.STRING, 'validWriteIdList', None, None, ), # 7
)
- def __init__(self, dbName=None, tblName=None, parts=None, ifNotExists=None, needResult=thrift_spec[5][4], catName=None, txnId=thrift_spec[7][4], validWriteIdList=None,):
+ def __init__(self, dbName=None, tblName=None, parts=None, ifNotExists=None, needResult=thrift_spec[5][4], catName=None, validWriteIdList=None,):
self.dbName = dbName
self.tblName = tblName
self.parts = parts
self.ifNotExists = ifNotExists
self.needResult = needResult
self.catName = catName
- self.txnId = txnId
self.validWriteIdList = validWriteIdList
def read(self, iprot):
@@ -10194,11 +10139,6 @@ class AddPartitionsRequest:
else:
iprot.skip(ftype)
elif fid == 7:
- if ftype == TType.I64:
- self.txnId = iprot.readI64()
- else:
- iprot.skip(ftype)
- elif fid == 8:
if ftype == TType.STRING:
self.validWriteIdList = iprot.readString()
else:
@@ -10240,12 +10180,8 @@ class AddPartitionsRequest:
oprot.writeFieldBegin('catName', TType.STRING, 6)
oprot.writeString(self.catName)
oprot.writeFieldEnd()
- if self.txnId is not None:
- oprot.writeFieldBegin('txnId', TType.I64, 7)
- oprot.writeI64(self.txnId)
- oprot.writeFieldEnd()
if self.validWriteIdList is not None:
- oprot.writeFieldBegin('validWriteIdList', TType.STRING, 8)
+ oprot.writeFieldBegin('validWriteIdList', TType.STRING, 7)
oprot.writeString(self.validWriteIdList)
oprot.writeFieldEnd()
oprot.writeFieldStop()
@@ -10271,7 +10207,6 @@ class AddPartitionsRequest:
value = (value * 31) ^ hash(self.ifNotExists)
value = (value * 31) ^ hash(self.needResult)
value = (value * 31) ^ hash(self.catName)
- value = (value * 31) ^ hash(self.txnId)
value = (value * 31) ^ hash(self.validWriteIdList)
return value
@@ -17526,7 +17461,6 @@ class GetTableRequest:
- tblName
- capabilities
- catName
- - txnId
- validWriteIdList
"""
@@ -17536,16 +17470,15 @@ class GetTableRequest:
(2, TType.STRING, 'tblName', None, None, ), # 2
(3, TType.STRUCT, 'capabilities', (ClientCapabilities, ClientCapabilities.thrift_spec), None, ), # 3
(4, TType.STRING, 'catName', None, None, ), # 4
- (5, TType.I64, 'txnId', None, -1, ), # 5
+ None, # 5
(6, TType.STRING, 'validWriteIdList', None, None, ), # 6
)
- def __init__(self, dbName=None, tblName=None, capabilities=None, catName=None, txnId=thrift_spec[5][4], validWriteIdList=None,):
+ def __init__(self, dbName=None, tblName=None, capabilities=None, catName=None, validWriteIdList=None,):
self.dbName = dbName
self.tblName = tblName
self.capabilities = capabilities
self.catName = catName
- self.txnId = txnId
self.validWriteIdList = validWriteIdList
def read(self, iprot):
@@ -17578,11 +17511,6 @@ class GetTableRequest:
self.catName = iprot.readString()
else:
iprot.skip(ftype)
- elif fid == 5:
- if ftype == TType.I64:
- self.txnId = iprot.readI64()
- else:
- iprot.skip(ftype)
elif fid == 6:
if ftype == TType.STRING:
self.validWriteIdList = iprot.readString()
@@ -17614,10 +17542,6 @@ class GetTableRequest:
oprot.writeFieldBegin('catName', TType.STRING, 4)
oprot.writeString(self.catName)
oprot.writeFieldEnd()
- if self.txnId is not None:
- oprot.writeFieldBegin('txnId', TType.I64, 5)
- oprot.writeI64(self.txnId)
- oprot.writeFieldEnd()
if self.validWriteIdList is not None:
oprot.writeFieldBegin('validWriteIdList', TType.STRING, 6)
oprot.writeString(self.validWriteIdList)
@@ -17639,7 +17563,6 @@ class GetTableRequest:
value = (value * 31) ^ hash(self.tblName)
value = (value * 31) ^ hash(self.capabilities)
value = (value * 31) ^ hash(self.catName)
- value = (value * 31) ^ hash(self.txnId)
value = (value * 31) ^ hash(self.validWriteIdList)
return value
@@ -22583,7 +22506,6 @@ class AlterPartitionsRequest:
- tableName
- partitions
- environmentContext
- - txnId
- writeId
- validWriteIdList
"""
@@ -22595,18 +22517,16 @@ class AlterPartitionsRequest:
(3, TType.STRING, 'tableName', None, None, ), # 3
(4, TType.LIST, 'partitions', (TType.STRUCT,(Partition, Partition.thrift_spec)), None, ), # 4
(5, TType.STRUCT, 'environmentContext', (EnvironmentContext, EnvironmentContext.thrift_spec), None, ), # 5
- (6, TType.I64, 'txnId', None, -1, ), # 6
- (7, TType.I64, 'writeId', None, -1, ), # 7
- (8, TType.STRING, 'validWriteIdList', None, None, ), # 8
+ (6, TType.I64, 'writeId', None, -1, ), # 6
+ (7, TType.STRING, 'validWriteIdList', None, None, ), # 7
)
- def __init__(self, catName=None, dbName=None, tableName=None, partitions=None, environmentContext=None, txnId=thrift_spec[6][4], writeId=thrift_spec[7][4], validWriteIdList=None,):
+ def __init__(self, catName=None, dbName=None, tableName=None, partitions=None, environmentContext=None, writeId=thrift_spec[6][4], validWriteIdList=None,):
self.catName = catName
self.dbName = dbName
self.tableName = tableName
self.partitions = partitions
self.environmentContext = environmentContext
- self.txnId = txnId
self.writeId = writeId
self.validWriteIdList = validWriteIdList
@@ -22653,15 +22573,10 @@ class AlterPartitionsRequest:
iprot.skip(ftype)
elif fid == 6:
if ftype == TType.I64:
- self.txnId = iprot.readI64()
- else:
- iprot.skip(ftype)
- elif fid == 7:
- if ftype == TType.I64:
self.writeId = iprot.readI64()
else:
iprot.skip(ftype)
- elif fid == 8:
+ elif fid == 7:
if ftype == TType.STRING:
self.validWriteIdList = iprot.readString()
else:
@@ -22699,16 +22614,12 @@ class AlterPartitionsRequest:
oprot.writeFieldBegin('environmentContext', TType.STRUCT, 5)
self.environmentContext.write(oprot)
oprot.writeFieldEnd()
- if self.txnId is not None:
- oprot.writeFieldBegin('txnId', TType.I64, 6)
- oprot.writeI64(self.txnId)
- oprot.writeFieldEnd()
if self.writeId is not None:
- oprot.writeFieldBegin('writeId', TType.I64, 7)
+ oprot.writeFieldBegin('writeId', TType.I64, 6)
oprot.writeI64(self.writeId)
oprot.writeFieldEnd()
if self.validWriteIdList is not None:
- oprot.writeFieldBegin('validWriteIdList', TType.STRING, 8)
+ oprot.writeFieldBegin('validWriteIdList', TType.STRING, 7)
oprot.writeString(self.validWriteIdList)
oprot.writeFieldEnd()
oprot.writeFieldStop()
@@ -22731,7 +22642,6 @@ class AlterPartitionsRequest:
value = (value * 31) ^ hash(self.tableName)
value = (value * 31) ^ hash(self.partitions)
value = (value * 31) ^ hash(self.environmentContext)
- value = (value * 31) ^ hash(self.txnId)
value = (value * 31) ^ hash(self.writeId)
value = (value * 31) ^ hash(self.validWriteIdList)
return value
@@ -22801,7 +22711,6 @@ class RenamePartitionRequest:
- tableName
- partVals
- newPart
- - txnId
- validWriteIdList
"""
@@ -22812,17 +22721,15 @@ class RenamePartitionRequest:
(3, TType.STRING, 'tableName', None, None, ), # 3
(4, TType.LIST, 'partVals', (TType.STRING,None), None, ), # 4
(5, TType.STRUCT, 'newPart', (Partition, Partition.thrift_spec), None, ), # 5
- (6, TType.I64, 'txnId', None, -1, ), # 6
- (7, TType.STRING, 'validWriteIdList', None, None, ), # 7
+ (6, TType.STRING, 'validWriteIdList', None, None, ), # 6
)
- def __init__(self, catName=None, dbName=None, tableName=None, partVals=None, newPart=None, txnId=thrift_spec[6][4], validWriteIdList=None,):
+ def __init__(self, catName=None, dbName=None, tableName=None, partVals=None, newPart=None, validWriteIdList=None,):
self.catName = catName
self.dbName = dbName
self.tableName = tableName
self.partVals = partVals
self.newPart = newPart
- self.txnId = txnId
self.validWriteIdList = validWriteIdList
def read(self, iprot):
@@ -22866,11 +22773,6 @@ class RenamePartitionRequest:
else:
iprot.skip(ftype)
elif fid == 6:
- if ftype == TType.I64:
- self.txnId = iprot.readI64()
- else:
- iprot.skip(ftype)
- elif fid == 7:
if ftype == TType.STRING:
self.validWriteIdList = iprot.readString()
else:
@@ -22908,12 +22810,8 @@ class RenamePartitionRequest:
oprot.writeFieldBegin('newPart', TType.STRUCT, 5)
self.newPart.write(oprot)
oprot.writeFieldEnd()
- if self.txnId is not None:
- oprot.writeFieldBegin('txnId', TType.I64, 6)
- oprot.writeI64(self.txnId)
- oprot.writeFieldEnd()
if self.validWriteIdList is not None:
- oprot.writeFieldBegin('validWriteIdList', TType.STRING, 7)
+ oprot.writeFieldBegin('validWriteIdList', TType.STRING, 6)
oprot.writeString(self.validWriteIdList)
oprot.writeFieldEnd()
oprot.writeFieldStop()
@@ -22938,7 +22836,6 @@ class RenamePartitionRequest:
value = (value * 31) ^ hash(self.tableName)
value = (value * 31) ^ hash(self.partVals)
value = (value * 31) ^ hash(self.newPart)
- value = (value * 31) ^ hash(self.txnId)
value = (value * 31) ^ hash(self.validWriteIdList)
return value
@@ -23007,7 +22904,6 @@ class AlterTableRequest:
- tableName
- table
- environmentContext
- - txnId
- writeId
- validWriteIdList
"""
@@ -23019,18 +22915,16 @@ class AlterTableRequest:
(3, TType.STRING, 'tableName', None, None, ), # 3
(4, TType.STRUCT, 'table', (Table, Table.thrift_spec), None, ), # 4
(5, TType.STRUCT, 'environmentContext', (EnvironmentContext, EnvironmentContext.thrift_spec), None, ), # 5
- (6, TType.I64, 'txnId', None, -1, ), # 6
- (7, TType.I64, 'writeId', None, -1, ), # 7
- (8, TType.STRING, 'validWriteIdList', None, None, ), # 8
+ (6, TType.I64, 'writeId', None, -1, ), # 6
+ (7, TType.STRING, 'validWriteIdList', None, None, ), # 7
)
- def __init__(self, catName=None, dbName=None, tableName=None, table=None, environmentContext=None, txnId=thrift_spec[6][4], writeId=thrift_spec[7][4], validWriteIdList=None,):
+ def __init__(self, catName=None, dbName=None, tableName=None, table=None, environmentContext=None, writeId=thrift_spec[6][4], validWriteIdList=None,):
self.catName = catName
self.dbName = dbName
self.tableName = tableName
self.table = table
self.environmentContext = environmentContext
- self.txnId = txnId
self.writeId = writeId
self.validWriteIdList = validWriteIdList
@@ -23072,15 +22966,10 @@ class AlterTableRequest:
iprot.skip(ftype)
elif fid == 6:
if ftype == TType.I64:
- self.txnId = iprot.readI64()
- else:
- iprot.skip(ftype)
- elif fid == 7:
- if ftype == TType.I64:
self.writeId = iprot.readI64()
else:
iprot.skip(ftype)
- elif fid == 8:
+ elif fid == 7:
if ftype == TType.STRING:
self.validWriteIdList = iprot.readString()
else:
@@ -23115,16 +23004,12 @@ class AlterTableRequest:
oprot.writeFieldBegin('environmentContext', TType.STRUCT, 5)
self.environmentContext.write(oprot)
oprot.writeFieldEnd()
- if self.txnId is not None:
- oprot.writeFieldBegin('txnId', TType.I64, 6)
- oprot.writeI64(self.txnId)
- oprot.writeFieldEnd()
if self.writeId is not None:
- oprot.writeFieldBegin('writeId', TType.I64, 7)
+ oprot.writeFieldBegin('writeId', TType.I64, 6)
oprot.writeI64(self.writeId)
oprot.writeFieldEnd()
if self.validWriteIdList is not None:
- oprot.writeFieldBegin('validWriteIdList', TType.STRING, 8)
+ oprot.writeFieldBegin('validWriteIdList', TType.STRING, 7)
oprot.writeString(self.validWriteIdList)
oprot.writeFieldEnd()
oprot.writeFieldStop()
@@ -23147,7 +23032,6 @@ class AlterTableRequest:
value = (value * 31) ^ hash(self.tableName)
value = (value * 31) ^ hash(self.table)
value = (value * 31) ^ hash(self.environmentContext)
- value = (value * 31) ^ hash(self.txnId)
value = (value * 31) ^ hash(self.writeId)
value = (value * 31) ^ hash(self.validWriteIdList)
return value
http://git-wip-us.apache.org/repos/asf/hive/blob/cdb32a7f/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/hive_metastore_types.rb
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/hive_metastore_types.rb b/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/hive_metastore_types.rb
index 2e2392d..e0c6c02 100644
--- a/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/hive_metastore_types.rb
+++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/hive_metastore_types.rb
@@ -608,15 +608,13 @@ class TruncateTableRequest
DBNAME = 1
TABLENAME = 2
PARTNAMES = 3
- TXNID = 4
- WRITEID = 5
- VALIDWRITEIDLIST = 6
+ WRITEID = 4
+ VALIDWRITEIDLIST = 5
FIELDS = {
DBNAME => {:type => ::Thrift::Types::STRING, :name => 'dbName'},
TABLENAME => {:type => ::Thrift::Types::STRING, :name => 'tableName'},
PARTNAMES => {:type => ::Thrift::Types::LIST, :name => 'partNames', :element => {:type => ::Thrift::Types::STRING}, :optional => true},
- TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true},
WRITEID => {:type => ::Thrift::Types::I64, :name => 'writeId', :default => -1, :optional => true},
VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true}
}
@@ -1646,14 +1644,12 @@ class SetPartitionsStatsRequest
include ::Thrift::Struct, ::Thrift::Struct_Union
COLSTATS = 1
NEEDMERGE = 2
- TXNID = 3
- WRITEID = 4
- VALIDWRITEIDLIST = 5
+ WRITEID = 3
+ VALIDWRITEIDLIST = 4
FIELDS = {
COLSTATS => {:type => ::Thrift::Types::LIST, :name => 'colStats', :element => {:type => ::Thrift::Types::STRUCT, :class => ::ColumnStatistics}},
NEEDMERGE => {:type => ::Thrift::Types::BOOL, :name => 'needMerge', :optional => true},
- TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true},
WRITEID => {:type => ::Thrift::Types::I64, :name => 'writeId', :default => -1, :optional => true},
VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true}
}
@@ -2178,15 +2174,13 @@ class TableStatsRequest
TBLNAME = 2
COLNAMES = 3
CATNAME = 4
- TXNID = 5
- VALIDWRITEIDLIST = 6
+ VALIDWRITEIDLIST = 5
FIELDS = {
DBNAME => {:type => ::Thrift::Types::STRING, :name => 'dbName'},
TBLNAME => {:type => ::Thrift::Types::STRING, :name => 'tblName'},
COLNAMES => {:type => ::Thrift::Types::LIST, :name => 'colNames', :element => {:type => ::Thrift::Types::STRING}},
CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName', :optional => true},
- TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true},
VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true}
}
@@ -2208,8 +2202,7 @@ class PartitionsStatsRequest
COLNAMES = 3
PARTNAMES = 4
CATNAME = 5
- TXNID = 6
- VALIDWRITEIDLIST = 7
+ VALIDWRITEIDLIST = 6
FIELDS = {
DBNAME => {:type => ::Thrift::Types::STRING, :name => 'dbName'},
@@ -2217,7 +2210,6 @@ class PartitionsStatsRequest
COLNAMES => {:type => ::Thrift::Types::LIST, :name => 'colNames', :element => {:type => ::Thrift::Types::STRING}},
PARTNAMES => {:type => ::Thrift::Types::LIST, :name => 'partNames', :element => {:type => ::Thrift::Types::STRING}},
CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName', :optional => true},
- TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true},
VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true}
}
@@ -2259,8 +2251,7 @@ class AddPartitionsRequest
IFNOTEXISTS = 4
NEEDRESULT = 5
CATNAME = 6
- TXNID = 7
- VALIDWRITEIDLIST = 8
+ VALIDWRITEIDLIST = 7
FIELDS = {
DBNAME => {:type => ::Thrift::Types::STRING, :name => 'dbName'},
@@ -2269,7 +2260,6 @@ class AddPartitionsRequest
IFNOTEXISTS => {:type => ::Thrift::Types::BOOL, :name => 'ifNotExists'},
NEEDRESULT => {:type => ::Thrift::Types::BOOL, :name => 'needResult', :default => true, :optional => true},
CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName', :optional => true},
- TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true},
VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true}
}
@@ -3915,7 +3905,6 @@ class GetTableRequest
TBLNAME = 2
CAPABILITIES = 3
CATNAME = 4
- TXNID = 5
VALIDWRITEIDLIST = 6
FIELDS = {
@@ -3923,7 +3912,6 @@ class GetTableRequest
TBLNAME => {:type => ::Thrift::Types::STRING, :name => 'tblName'},
CAPABILITIES => {:type => ::Thrift::Types::STRUCT, :name => 'capabilities', :class => ::ClientCapabilities, :optional => true},
CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName', :optional => true},
- TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true},
VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true}
}
@@ -5114,9 +5102,8 @@ class AlterPartitionsRequest
TABLENAME = 3
PARTITIONS = 4
ENVIRONMENTCONTEXT = 5
- TXNID = 6
- WRITEID = 7
- VALIDWRITEIDLIST = 8
+ WRITEID = 6
+ VALIDWRITEIDLIST = 7
FIELDS = {
CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName', :optional => true},
@@ -5124,7 +5111,6 @@ class AlterPartitionsRequest
TABLENAME => {:type => ::Thrift::Types::STRING, :name => 'tableName'},
PARTITIONS => {:type => ::Thrift::Types::LIST, :name => 'partitions', :element => {:type => ::Thrift::Types::STRUCT, :class => ::Partition}},
ENVIRONMENTCONTEXT => {:type => ::Thrift::Types::STRUCT, :name => 'environmentContext', :class => ::EnvironmentContext, :optional => true},
- TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true},
WRITEID => {:type => ::Thrift::Types::I64, :name => 'writeId', :default => -1, :optional => true},
VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true}
}
@@ -5162,8 +5148,7 @@ class RenamePartitionRequest
TABLENAME = 3
PARTVALS = 4
NEWPART = 5
- TXNID = 6
- VALIDWRITEIDLIST = 7
+ VALIDWRITEIDLIST = 6
FIELDS = {
CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName', :optional => true},
@@ -5171,7 +5156,6 @@ class RenamePartitionRequest
TABLENAME => {:type => ::Thrift::Types::STRING, :name => 'tableName'},
PARTVALS => {:type => ::Thrift::Types::LIST, :name => 'partVals', :element => {:type => ::Thrift::Types::STRING}},
NEWPART => {:type => ::Thrift::Types::STRUCT, :name => 'newPart', :class => ::Partition},
- TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true},
VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true}
}
@@ -5209,9 +5193,8 @@ class AlterTableRequest
TABLENAME = 3
TABLE = 4
ENVIRONMENTCONTEXT = 5
- TXNID = 6
- WRITEID = 7
- VALIDWRITEIDLIST = 8
+ WRITEID = 6
+ VALIDWRITEIDLIST = 7
FIELDS = {
CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName', :optional => true},
@@ -5219,7 +5202,6 @@ class AlterTableRequest
TABLENAME => {:type => ::Thrift::Types::STRING, :name => 'tableName'},
TABLE => {:type => ::Thrift::Types::STRUCT, :name => 'table', :class => ::Table},
ENVIRONMENTCONTEXT => {:type => ::Thrift::Types::STRUCT, :name => 'environmentContext', :class => ::EnvironmentContext, :optional => true},
- TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true},
WRITEID => {:type => ::Thrift::Types::I64, :name => 'writeId', :default => -1, :optional => true},
VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true}
}
http://git-wip-us.apache.org/repos/asf/hive/blob/cdb32a7f/standalone-metastore/metastore-common/src/main/thrift/hive_metastore.thrift
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-common/src/main/thrift/hive_metastore.thrift b/standalone-metastore/metastore-common/src/main/thrift/hive_metastore.thrift
index 4cfa6a1..8a4bdd8 100644
--- a/standalone-metastore/metastore-common/src/main/thrift/hive_metastore.thrift
+++ b/standalone-metastore/metastore-common/src/main/thrift/hive_metastore.thrift
@@ -282,9 +282,8 @@ struct TruncateTableRequest {
1: required string dbName,
2: required string tableName,
3: optional list<string> partNames,
- 4: optional i64 txnId=-1,
- 5: optional i64 writeId=-1,
- 6: optional string validWriteIdList
+ 4: optional i64 writeId=-1,
+ 5: optional string validWriteIdList
}
struct TruncateTableResponse {
@@ -596,9 +595,8 @@ struct AggrStats {
struct SetPartitionsStatsRequest {
1: required list<ColumnStatistics> colStats,
2: optional bool needMerge, //stats need to be merged with the existing stats
-3: optional i64 txnId=-1, // transaction id of the query that sends this structure
-4: optional i64 writeId=-1, // writeId for the current query that updates the stats
-5: optional string validWriteIdList // valid write id list for the table for which this struct is being sent
+3: optional i64 writeId=-1, // writeId for the current query that updates the stats
+4: optional string validWriteIdList // valid write id list for the table for which this struct is being sent
}
struct SetPartitionsStatsResponse {
@@ -745,8 +743,7 @@ struct TableStatsRequest {
2: required string tblName,
3: required list<string> colNames
4: optional string catName,
- 5: optional i64 txnId=-1, // transaction id of the query that sends this structure
- 6: optional string validWriteIdList // valid write id list for the table for which this struct is being sent
+ 5: optional string validWriteIdList // valid write id list for the table for which this struct is being sent
}
struct PartitionsStatsRequest {
@@ -755,8 +752,7 @@ struct PartitionsStatsRequest {
3: required list<string> colNames,
4: required list<string> partNames,
5: optional string catName,
- 6: optional i64 txnId=-1, // transaction id of the query that sends this structure
- 7: optional string validWriteIdList // valid write id list for the table for which this struct is being sent
+ 6: optional string validWriteIdList // valid write id list for the table for which this struct is being sent
}
// Return type for add_partitions_req
@@ -773,8 +769,7 @@ struct AddPartitionsRequest {
4: required bool ifNotExists,
5: optional bool needResult=true,
6: optional string catName,
- 7: optional i64 txnId=-1,
- 8: optional string validWriteIdList
+ 7: optional string validWriteIdList
}
// Return type for drop_partitions_req
@@ -1275,7 +1270,6 @@ struct GetTableRequest {
2: required string tblName,
3: optional ClientCapabilities capabilities,
4: optional string catName,
- 5: optional i64 txnId=-1,
6: optional string validWriteIdList
}
@@ -1615,9 +1609,8 @@ struct AlterPartitionsRequest {
3: required string tableName,
4: required list<Partition> partitions,
5: optional EnvironmentContext environmentContext,
- 6: optional i64 txnId=-1,
- 7: optional i64 writeId=-1,
- 8: optional string validWriteIdList
+ 6: optional i64 writeId=-1,
+ 7: optional string validWriteIdList
}
struct AlterPartitionsResponse {
@@ -1629,8 +1622,7 @@ struct RenamePartitionRequest {
3: required string tableName,
4: required list<string> partVals,
5: required Partition newPart,
- 6: optional i64 txnId=-1,
- 7: optional string validWriteIdList
+ 6: optional string validWriteIdList
}
struct RenamePartitionResponse {
@@ -1642,9 +1634,8 @@ struct AlterTableRequest {
3: required string tableName,
4: required Table table,
5: optional EnvironmentContext environmentContext,
- 6: optional i64 txnId=-1,
- 7: optional i64 writeId=-1,
- 8: optional string validWriteIdList
+ 6: optional i64 writeId=-1,
+ 7: optional string validWriteIdList
// TODO: also add cascade here, out of envCtx
}
http://git-wip-us.apache.org/repos/asf/hive/blob/cdb32a7f/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/AlterHandler.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/AlterHandler.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/AlterHandler.java
index f3dc264..f4e0c41 100644
--- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/AlterHandler.java
+++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/AlterHandler.java
@@ -61,7 +61,7 @@ public interface AlterHandler extends Configurable {
default void alterTable(RawStore msdb, Warehouse wh, String catName, String dbname,
String name, Table newTable, EnvironmentContext envContext)
throws InvalidOperationException, MetaException {
- alterTable(msdb, wh, catName, dbname, name, newTable, envContext, null, -1, null);
+ alterTable(msdb, wh, catName, dbname, name, newTable, envContext, null, null);
}
/**
@@ -88,7 +88,7 @@ public interface AlterHandler extends Configurable {
*/
void alterTable(RawStore msdb, Warehouse wh, String catName, String dbname,
String name, Table newTable, EnvironmentContext envContext,
- IHMSHandler handler, long txnId, String writeIdList)
+ IHMSHandler handler, String writeIdList)
throws InvalidOperationException, MetaException;
/**
@@ -146,7 +146,7 @@ public interface AlterHandler extends Configurable {
Partition alterPartition(final RawStore msdb, Warehouse wh, final String catName,
final String dbname, final String name, final List<String> part_vals,
final Partition new_part, EnvironmentContext environmentContext,
- IHMSHandler handler, long txnId, String validWriteIds)
+ IHMSHandler handler, String validWriteIds)
throws InvalidOperationException, InvalidObjectException, AlreadyExistsException, MetaException;
/**
@@ -198,7 +198,7 @@ public interface AlterHandler extends Configurable {
*/
List<Partition> alterPartitions(final RawStore msdb, Warehouse wh, final String catName,
final String dbname, final String name, final List<Partition> new_parts,
- EnvironmentContext environmentContext, long txnId, String writeIdList, long writeId,
+ EnvironmentContext environmentContext, String writeIdList, long writeId,
IHMSHandler handler)
throws InvalidOperationException, InvalidObjectException, AlreadyExistsException, MetaException;
}
\ No newline at end of file
http://git-wip-us.apache.org/repos/asf/hive/blob/cdb32a7f/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java
index 719f001..e16bf7a 100644
--- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java
+++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java
@@ -90,7 +90,7 @@ public class HiveAlterHandler implements AlterHandler {
@Override
public void alterTable(RawStore msdb, Warehouse wh, String catName, String dbname,
String name, Table newt, EnvironmentContext environmentContext,
- IHMSHandler handler, long txnId, String writeIdList)
+ IHMSHandler handler, String writeIdList)
throws InvalidOperationException, MetaException {
catName = normalizeIdentifier(catName);
name = name.toLowerCase();
@@ -142,7 +142,7 @@ public class HiveAlterHandler implements AlterHandler {
// check if table with the new name already exists
if (!newTblName.equals(name) || !newDbName.equals(dbname)) {
- if (msdb.getTable(catName, newDbName, newTblName, -1, null) != null) {
+ if (msdb.getTable(catName, newDbName, newTblName, null) != null) {
throw new InvalidOperationException("new table " + newDbName
+ "." + newTblName + " already exists");
}
@@ -152,7 +152,7 @@ public class HiveAlterHandler implements AlterHandler {
msdb.openTransaction();
// get old table
// Note: we don't verify stats here; it's done below in alterTableUpdateTableColumnStats.
- oldt = msdb.getTable(catName, dbname, name, -1, null);
+ oldt = msdb.getTable(catName, dbname, name, null);
if (oldt == null) {
throw new InvalidOperationException("table " +
TableName.getQualified(catName, dbname, name) + " doesn't exist");
@@ -278,7 +278,7 @@ public class HiveAlterHandler implements AlterHandler {
}
}
// Do not verify stats parameters on a partitioned table.
- msdb.alterTable(catName, dbname, name, newt, -1, null);
+ msdb.alterTable(catName, dbname, name, newt, null);
// alterPartition is only for changing the partition location in the table rename
if (dataWasMoved) {
@@ -297,7 +297,7 @@ public class HiveAlterHandler implements AlterHandler {
partValues.add(part.getValues());
}
msdb.alterPartitions(catName, newDbName, newTblName, partValues,
- partBatch, newt.getWriteId(), txnId, writeIdList);
+ partBatch, newt.getWriteId(), writeIdList);
}
}
@@ -306,11 +306,11 @@ public class HiveAlterHandler implements AlterHandler {
newPartColStats.getStatsDesc().setDbName(newDbName);
newPartColStats.getStatsDesc().setTableName(newTblName);
msdb.updatePartitionColumnStatistics(newPartColStats, partColStats.getKey().getValues(),
- txnId, writeIdList, newt.getWriteId());
+ writeIdList, newt.getWriteId());
}
} else {
alterTableUpdateTableColumnStats(
- msdb, oldt, newt, environmentContext, txnId, writeIdList);
+ msdb, oldt, newt, environmentContext, writeIdList);
}
} else {
// operations other than table rename
@@ -334,24 +334,24 @@ public class HiveAlterHandler implements AlterHandler {
assert(colStats == null);
if (cascade) {
msdb.alterPartition(
- catName, dbname, name, part.getValues(), part, txnId, writeIdList);
+ catName, dbname, name, part.getValues(), part, writeIdList);
} else {
// update changed properties (stats)
oldPart.setParameters(part.getParameters());
msdb.alterPartition(
- catName, dbname, name, part.getValues(), oldPart, txnId, writeIdList);
+ catName, dbname, name, part.getValues(), oldPart, writeIdList);
}
}
// Don't validate table-level stats for a partitoned table.
- msdb.alterTable(catName, dbname, name, newt, -1, null);
+ msdb.alterTable(catName, dbname, name, newt, null);
} else {
LOG.warn("Alter table not cascaded to partitions.");
alterTableUpdateTableColumnStats(
- msdb, oldt, newt, environmentContext, txnId, writeIdList);
+ msdb, oldt, newt, environmentContext, writeIdList);
}
} else {
alterTableUpdateTableColumnStats(
- msdb, oldt, newt, environmentContext, txnId, writeIdList);
+ msdb, oldt, newt, environmentContext, writeIdList);
}
}
@@ -431,13 +431,13 @@ public class HiveAlterHandler implements AlterHandler {
EnvironmentContext environmentContext)
throws InvalidOperationException, InvalidObjectException, AlreadyExistsException, MetaException {
return alterPartition(msdb, wh, DEFAULT_CATALOG_NAME, dbname, name, part_vals, new_part,
- environmentContext, null, -1, null);
+ environmentContext, null, null);
}
@Override
public Partition alterPartition(RawStore msdb, Warehouse wh, String catName, String dbname,
String name, List<String> part_vals, final Partition new_part,
- EnvironmentContext environmentContext, IHMSHandler handler, long txnId, String validWriteIds)
+ EnvironmentContext environmentContext, IHMSHandler handler, String validWriteIds)
throws InvalidOperationException, InvalidObjectException, AlreadyExistsException, MetaException {
boolean success = false;
Partition oldPart;
@@ -459,7 +459,7 @@ public class HiveAlterHandler implements AlterHandler {
try {
msdb.openTransaction();
- Table tbl = msdb.getTable(catName, dbname, name, -1, null);
+ Table tbl = msdb.getTable(catName, dbname, name, null);
if (tbl == null) {
throw new InvalidObjectException(
"Unable to alter partition because table or database does not exist.");
@@ -481,7 +481,7 @@ public class HiveAlterHandler implements AlterHandler {
oldPart.getSd().getCols(), tbl, new_part, null);
}
msdb.alterPartition(
- catName, dbname, name, new_part.getValues(), new_part, txnId, validWriteIds);
+ catName, dbname, name, new_part.getValues(), new_part, validWriteIds);
if (transactionalListeners != null && !transactionalListeners.isEmpty()) {
MetaStoreListenerNotifier.notifyEvent(transactionalListeners,
EventMessage.EventType.ALTER_PARTITION,
@@ -516,7 +516,7 @@ public class HiveAlterHandler implements AlterHandler {
Database db;
try {
msdb.openTransaction();
- Table tbl = msdb.getTable(DEFAULT_CATALOG_NAME, dbname, name, -1, null);
+ Table tbl = msdb.getTable(DEFAULT_CATALOG_NAME, dbname, name, null);
if (tbl == null) {
throw new InvalidObjectException(
"Unable to alter partition because table or database does not exist.");
@@ -617,12 +617,12 @@ public class HiveAlterHandler implements AlterHandler {
String newPartName = Warehouse.makePartName(tbl.getPartitionKeys(), new_part.getValues());
ColumnStatistics cs = updateOrGetPartitionColumnStats(msdb, catName, dbname, name, oldPart.getValues(),
oldPart.getSd().getCols(), tbl, new_part, null);
- msdb.alterPartition(catName, dbname, name, part_vals, new_part, txnId, validWriteIds);
+ msdb.alterPartition(catName, dbname, name, part_vals, new_part, validWriteIds);
if (cs != null) {
cs.getStatsDesc().setPartName(newPartName);
try {
msdb.updatePartitionColumnStatistics(cs, new_part.getValues(),
- txnId, validWriteIds, new_part.getWriteId());
+ validWriteIds, new_part.getWriteId());
} catch (InvalidInputException iie) {
throw new InvalidOperationException("Unable to update partition stats in table rename." + iie);
} catch (NoSuchObjectException nsoe) {
@@ -668,7 +668,7 @@ public class HiveAlterHandler implements AlterHandler {
EnvironmentContext environmentContext)
throws InvalidOperationException, InvalidObjectException, AlreadyExistsException, MetaException {
return alterPartitions(msdb, wh, DEFAULT_CATALOG_NAME, dbname, name, new_parts,
- environmentContext, -1, null, -1, null);
+ environmentContext, null, -1, null);
}
@Override
@@ -676,7 +676,7 @@ public class HiveAlterHandler implements AlterHandler {
final String dbname, final String name,
final List<Partition> new_parts,
EnvironmentContext environmentContext,
- long txnId, String writeIdList, long writeId,
+ String writeIdList, long writeId,
IHMSHandler handler)
throws InvalidOperationException, InvalidObjectException, AlreadyExistsException, MetaException {
List<Partition> oldParts = new ArrayList<>();
@@ -691,7 +691,7 @@ public class HiveAlterHandler implements AlterHandler {
msdb.openTransaction();
// Note: should we pass in write ID here? We only update stats on parts so probably not.
- Table tbl = msdb.getTable(catName, dbname, name, -1, null);
+ Table tbl = msdb.getTable(catName, dbname, name, null);
if (tbl == null) {
throw new InvalidObjectException(
"Unable to alter partitions because table or database does not exist.");
@@ -726,7 +726,7 @@ public class HiveAlterHandler implements AlterHandler {
}
}
- msdb.alterPartitions(catName, dbname, name, partValsList, new_parts, writeId, txnId, writeIdList);
+ msdb.alterPartitions(catName, dbname, name, partValsList, new_parts, writeId, writeIdList);
Iterator<Partition> oldPartsIt = oldParts.iterator();
for (Partition newPart : new_parts) {
Partition oldPart;
@@ -795,7 +795,7 @@ public class HiveAlterHandler implements AlterHandler {
@VisibleForTesting
void alterTableUpdateTableColumnStats(RawStore msdb, Table oldTable, Table newTable,
- EnvironmentContext ec, long txnId, String validWriteIds)
+ EnvironmentContext ec, String validWriteIds)
throws MetaException, InvalidObjectException {
String catName = normalizeIdentifier(oldTable.isSetCatName() ? oldTable.getCatName() :
getDefaultCatalog(conf));
@@ -854,13 +854,13 @@ public class HiveAlterHandler implements AlterHandler {
}
// Change to new table and append stats for the new table
- msdb.alterTable(catName, dbName, tableName, newTable, txnId, validWriteIds);
+ msdb.alterTable(catName, dbName, tableName, newTable, validWriteIds);
if (updateColumnStats && !newStatsObjs.isEmpty()) {
ColumnStatisticsDesc statsDesc = colStats.getStatsDesc();
statsDesc.setDbName(newDbName);
statsDesc.setTableName(newTableName);
colStats.setStatsObj(newStatsObjs);
- msdb.updateTableColumnStatistics(colStats, txnId, validWriteIds, newTable.getWriteId());
+ msdb.updateTableColumnStatistics(colStats, validWriteIds, newTable.getWriteId());
}
} catch (NoSuchObjectException nsoe) {
LOG.debug("Could not find db entry." + nsoe);
http://git-wip-us.apache.org/repos/asf/hive/blob/cdb32a7f/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java
index 285f7fb..eb4eb1b 100644
--- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java
+++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java
@@ -2435,7 +2435,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
private boolean is_table_exists(RawStore ms, String catName, String dbname, String name)
throws MetaException {
- return (ms.getTable(catName, dbname, name, -1, null) != null);
+ return (ms.getTable(catName, dbname, name, null) != null);
}
private boolean drop_table_core(final RawStore ms, final String catName, final String dbname,
@@ -2694,7 +2694,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
}
private void alterPartitionForTruncate(RawStore ms, String catName, String dbName, String tableName,
- Table table, Partition partition, long txnId, String validWriteIds, long writeId) throws Exception {
+ Table table, Partition partition, String validWriteIds, long writeId) throws Exception {
EnvironmentContext environmentContext = new EnvironmentContext();
updateStatsForTruncate(partition.getParameters(), environmentContext);
@@ -2714,17 +2714,17 @@ public class HiveMetaStore extends ThriftHiveMetastore {
partition.setWriteId(writeId);
}
alterHandler.alterPartition(ms, wh, catName, dbName, tableName, null, partition,
- environmentContext, this, txnId, validWriteIds);
+ environmentContext, this, validWriteIds);
}
private void alterTableStatsForTruncate(RawStore ms, String catName, String dbName,
String tableName, Table table, List<String> partNames,
- long txnId, String validWriteIds, long writeId) throws Exception {
+ String validWriteIds, long writeId) throws Exception {
if (partNames == null) {
if (0 != table.getPartitionKeysSize()) {
for (Partition partition : ms.getPartitions(catName, dbName, tableName, Integer.MAX_VALUE)) {
alterPartitionForTruncate(ms, catName, dbName, tableName, table, partition,
- txnId, validWriteIds, writeId);
+ validWriteIds, writeId);
}
} else {
EnvironmentContext environmentContext = new EnvironmentContext();
@@ -2747,12 +2747,12 @@ public class HiveMetaStore extends ThriftHiveMetastore {
table.setWriteId(writeId);
}
alterHandler.alterTable(ms, wh, catName, dbName, tableName, table,
- environmentContext, this, txnId, validWriteIds);
+ environmentContext, this, validWriteIds);
}
} else {
for (Partition partition : ms.getPartitionsByNames(catName, dbName, tableName, partNames)) {
alterPartitionForTruncate(ms, catName, dbName, tableName, table, partition,
- txnId, validWriteIds, writeId);
+ validWriteIds, writeId);
}
}
return;
@@ -2791,19 +2791,19 @@ public class HiveMetaStore extends ThriftHiveMetastore {
public void truncate_table(final String dbName, final String tableName, List<String> partNames)
throws NoSuchObjectException, MetaException {
// Deprecated path, won't work for txn tables.
- truncateTableInternal(dbName, tableName, partNames, -1, null, -1);
+ truncateTableInternal(dbName, tableName, partNames, null, -1);
}
@Override
public TruncateTableResponse truncate_table_req(TruncateTableRequest req)
throws MetaException, TException {
truncateTableInternal(req.getDbName(), req.getTableName(), req.getPartNames(),
- req.getTxnId(), req.getValidWriteIdList(), req.getWriteId());
+ req.getValidWriteIdList(), req.getWriteId());
return new TruncateTableResponse();
}
private void truncateTableInternal(String dbName, String tableName, List<String> partNames,
- long txnId, String validWriteIds, long writeId) throws MetaException, NoSuchObjectException {
+ String validWriteIds, long writeId) throws MetaException, NoSuchObjectException {
try {
String[] parsedDbName = parseDbName(dbName, conf);
Table tbl = get_table_core(parsedDbName[CAT_NAME], parsedDbName[DB_NAME], tableName);
@@ -2835,7 +2835,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
// Alter the table/partition stats and also notify truncate table event
alterTableStatsForTruncate(getMS(), parsedDbName[CAT_NAME], parsedDbName[DB_NAME],
- tableName, tbl, partNames, txnId, validWriteIds, writeId);
+ tableName, tbl, partNames, validWriteIds, writeId);
} catch (IOException e) {
throw new MetaException(e.getMessage());
} catch (MetaException | NoSuchObjectException e) {
@@ -2866,7 +2866,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
NoSuchObjectException {
String[] parsedDbName = parseDbName(dbname, conf);
return getTableInternal(
- parsedDbName[CAT_NAME], parsedDbName[DB_NAME], name, null, -1, null);
+ parsedDbName[CAT_NAME], parsedDbName[DB_NAME], name, null, null);
}
@Override
@@ -2874,11 +2874,11 @@ public class HiveMetaStore extends ThriftHiveMetastore {
NoSuchObjectException {
String catName = req.isSetCatName() ? req.getCatName() : getDefaultCatalog(conf);
return new GetTableResult(getTableInternal(catName, req.getDbName(), req.getTblName(),
- req.getCapabilities(), req.getTxnId(), req.getValidWriteIdList()));
+ req.getCapabilities(), req.getValidWriteIdList()));
}
private Table getTableInternal(String catName, String dbname, String name,
- ClientCapabilities capabilities, long txnId, String writeIdList)
+ ClientCapabilities capabilities, String writeIdList)
throws MetaException, NoSuchObjectException {
if (isInTest) {
assertClientHasCapability(capabilities, ClientCapability.TEST_CAPABILITY,
@@ -2889,7 +2889,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
startTableFunction("get_table", catName, dbname, name);
Exception ex = null;
try {
- t = get_table_core(catName, dbname, name, txnId, writeIdList);
+ t = get_table_core(catName, dbname, name, writeIdList);
if (MetaStoreUtils.isInsertOnlyTableParam(t.getParameters())) {
assertClientHasCapability(capabilities, ClientCapability.INSERT_ONLY_TABLES,
"insert-only tables", "get_table_req");
@@ -2929,7 +2929,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
final String dbname,
final String name)
throws MetaException, NoSuchObjectException {
- return get_table_core(catName, dbname, name, -1, null);
+ return get_table_core(catName, dbname, name, null);
}
@Override
@@ -2937,12 +2937,11 @@ public class HiveMetaStore extends ThriftHiveMetastore {
final String catName,
final String dbname,
final String name,
- final long txnId,
final String writeIdList)
throws MetaException, NoSuchObjectException {
Table t = null;
try {
- t = getMS().getTable(catName, dbname, name, txnId, writeIdList);
+ t = getMS().getTable(catName, dbname, name, writeIdList);
if (t == null) {
throw new NoSuchObjectException(TableName.getQualified(catName, dbname, name) +
" table not found");
@@ -3124,7 +3123,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
MetaStoreUtils.validatePartitionNameCharacters(part_vals, partitionValidationPattern);
- tbl = ms.getTable(part.getCatName(), part.getDbName(), part.getTableName(), -1, null);
+ tbl = ms.getTable(part.getCatName(), part.getDbName(), part.getTableName(), null);
if (tbl == null) {
throw new InvalidObjectException(
"Unable to add partition because table or database do not exist");
@@ -3318,7 +3317,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
try {
ms.openTransaction();
- tbl = ms.getTable(catName, dbName, tblName, -1, null);
+ tbl = ms.getTable(catName, dbName, tblName, null);
if (tbl == null) {
throw new InvalidObjectException("Unable to add partitions because "
+ TableName.getQualified(catName, dbName, tblName) +
@@ -3674,7 +3673,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
Database db = null;
try {
ms.openTransaction();
- tbl = ms.getTable(catName, dbName, tblName, -1, null);
+ tbl = ms.getTable(catName, dbName, tblName, null);
if (tbl == null) {
throw new InvalidObjectException("Unable to add partitions because "
+ "database or table " + dbName + "." + tblName + " does not exist");
@@ -3844,7 +3843,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
}
try {
ms.openTransaction();
- tbl = ms.getTable(part.getCatName(), part.getDbName(), part.getTableName(), -1, null);
+ tbl = ms.getTable(part.getCatName(), part.getDbName(), part.getTableName(), null);
if (tbl == null) {
throw new InvalidObjectException(
"Unable to add partition because table or database do not exist");
@@ -3966,7 +3965,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
Table destinationTable =
ms.getTable(
- parsedDestDbName[CAT_NAME], parsedDestDbName[DB_NAME], destTableName, -1, null);
+ parsedDestDbName[CAT_NAME], parsedDestDbName[DB_NAME], destTableName, null);
if (destinationTable == null) {
throw new MetaException( "The destination table " +
TableName.getQualified(parsedDestDbName[CAT_NAME],
@@ -3974,7 +3973,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
}
Table sourceTable =
ms.getTable(
- parsedSourceDbName[CAT_NAME], parsedSourceDbName[DB_NAME], sourceTableName, -1, null);
+ parsedSourceDbName[CAT_NAME], parsedSourceDbName[DB_NAME], sourceTableName, null);
if (sourceTable == null) {
throw new MetaException("The source table " +
TableName.getQualified(parsedSourceDbName[CAT_NAME],
@@ -4151,7 +4150,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
try {
ms.openTransaction();
part = ms.getPartition(catName, db_name, tbl_name, part_vals);
- tbl = get_table_core(catName, db_name, tbl_name, -1, null);
+ tbl = get_table_core(catName, db_name, tbl_name, null);
tableDataShouldBeDeleted = checkTableDataShouldBeDeleted(tbl, deleteData);
firePreEvent(new PreDropPartitionEvent(tbl, part, deleteData, this));
mustPurge = isMustPurge(envContext, tbl);
@@ -4826,7 +4825,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
String[] parsedDbName = parseDbName(dbName, conf);
// TODO: this method name is confusing, it actually does full alter (sortof)
rename_partition(parsedDbName[CAT_NAME], parsedDbName[DB_NAME], tableName, null, newPartition,
- envContext, -1, null);
+ envContext, null);
}
@Deprecated
@@ -4837,18 +4836,18 @@ public class HiveMetaStore extends ThriftHiveMetastore {
// Call rename_partition without an environment context.
String[] parsedDbName = parseDbName(db_name, conf);
rename_partition(parsedDbName[CAT_NAME], parsedDbName[DB_NAME], tbl_name, part_vals, new_part,
- null, -1, null);
+ null, null);
}
public RenamePartitionResponse rename_partition_req(
RenamePartitionRequest req) throws InvalidOperationException ,MetaException ,TException {
rename_partition(req.getCatName(), req.getDbName(), req.getTableName(), req.getPartVals(),
- req.getNewPart(), null, req.getTxnId(), req.getValidWriteIdList());
+ req.getNewPart(), null, req.getValidWriteIdList());
return new RenamePartitionResponse();
};
private void rename_partition(String catName, String db_name, String tbl_name,
- List<String> part_vals, Partition new_part, EnvironmentContext envContext, long txnId,
+ List<String> part_vals, Partition new_part, EnvironmentContext envContext,
String validWriteIds) throws TException {
startTableFunction("alter_partition", catName, db_name, tbl_name);
@@ -4883,13 +4882,13 @@ public class HiveMetaStore extends ThriftHiveMetastore {
}
oldPart = alterHandler.alterPartition(getMS(), wh, catName, db_name, tbl_name,
- part_vals, new_part, envContext, this, txnId, validWriteIds);
+ part_vals, new_part, envContext, this, validWriteIds);
// Only fetch the table if we actually have a listener
Table table = null;
if (!listeners.isEmpty()) {
if (table == null) {
- table = getMS().getTable(catName, db_name, tbl_name, -1, null);
+ table = getMS().getTable(catName, db_name, tbl_name, null);
}
MetaStoreListenerNotifier.notifyEvent(listeners,
@@ -4920,14 +4919,13 @@ public class HiveMetaStore extends ThriftHiveMetastore {
throws TException {
String[] o = parseDbName(db_name, conf);
alter_partitions_with_environment_context(o[0], o[1],
- tbl_name, new_parts, null, -1, null, -1);
+ tbl_name, new_parts, null, null, -1);
}
@Override
public AlterPartitionsResponse alter_partitions_req(AlterPartitionsRequest req) throws TException {
alter_partitions_with_environment_context(req.getCatName(),
req.getDbName(), req.getTableName(), req.getPartitions(), req.getEnvironmentContext(),
- req.isSetTxnId() ? req.getTxnId() : -1,
req.isSetValidWriteIdList() ? req.getValidWriteIdList() : null,
req.isSetWriteId() ? req.getWriteId() : -1);
return new AlterPartitionsResponse();
@@ -4941,12 +4939,12 @@ public class HiveMetaStore extends ThriftHiveMetastore {
throws TException {
String[] o = parseDbName(db_name, conf);
alter_partitions_with_environment_context(o[0], o[1], tbl_name, new_parts, environmentContext,
- -1, null, -1);
+ null, -1);
}
private void alter_partitions_with_environment_context(String catName, String db_name, final String tbl_name,
final List<Partition> new_parts, EnvironmentContext environmentContext,
- long txnId, String writeIdList, long writeId)
+ String writeIdList, long writeId)
throws TException {
if (environmentContext == null) {
environmentContext = new EnvironmentContext();
@@ -4975,7 +4973,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
firePreEvent(new PreAlterPartitionEvent(db_name, tbl_name, null, tmpPart, this));
}
oldParts = alterHandler.alterPartitions(getMS(), wh,
- catName, db_name, tbl_name, new_parts, environmentContext, txnId, writeIdList, writeId, this);
+ catName, db_name, tbl_name, new_parts, environmentContext, writeIdList, writeId, this);
Iterator<Partition> olditr = oldParts.iterator();
// Only fetch the table if we have a listener that needs it.
Table table = null;
@@ -4989,7 +4987,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
}
if (table == null) {
- table = getMS().getTable(catName, db_name, tbl_name, -1, null);
+ table = getMS().getTable(catName, db_name, tbl_name, null);
}
if (!listeners.isEmpty()) {
@@ -5028,7 +5026,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
// Do not set an environment context.
String[] parsedDbName = parseDbName(dbname, conf);
alter_table_core(parsedDbName[CAT_NAME], parsedDbName[DB_NAME], name, newTable,
- null, -1, null);
+ null, null);
}
@Override
@@ -5042,14 +5040,14 @@ public class HiveMetaStore extends ThriftHiveMetastore {
}
String[] parsedDbName = parseDbName(dbname, conf);
alter_table_core(parsedDbName[CAT_NAME], parsedDbName[DB_NAME], name, newTable,
- envContext, -1, null);
+ envContext, null);
}
@Override
public AlterTableResponse alter_table_req(AlterTableRequest req)
throws InvalidOperationException, MetaException, TException {
alter_table_core(req.getCatName(), req.getDbName(), req.getTableName(),
- req.getTable(), req.getEnvironmentContext(), req.getTxnId(), req.getValidWriteIdList());
+ req.getTable(), req.getEnvironmentContext(), req.getValidWriteIdList());
return new AlterTableResponse();
}
@@ -5060,11 +5058,11 @@ public class HiveMetaStore extends ThriftHiveMetastore {
throws InvalidOperationException, MetaException {
String[] parsedDbName = parseDbName(dbname, conf);
alter_table_core(parsedDbName[CAT_NAME], parsedDbName[DB_NAME],
- name, newTable, envContext, -1, null);
+ name, newTable, envContext, null);
}
private void alter_table_core(String catName, String dbname, String name, Table newTable,
- EnvironmentContext envContext, long txnId, String validWriteIdList)
+ EnvironmentContext envContext, String validWriteIdList)
throws InvalidOperationException, MetaException {
startFunction("alter_table", ": " + TableName.getQualified(catName, dbname, name)
+ " newtbl=" + newTable.getTableName());
@@ -5101,7 +5099,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
Table oldt = get_table_core(catName, dbname, name);
firePreEvent(new PreAlterTableEvent(oldt, newTable, this));
alterHandler.alterTable(getMS(), wh, catName, dbname, name, newTable,
- envContext, this, txnId, validWriteIdList);
+ envContext, this, validWriteIdList);
success = true;
} catch (NoSuchObjectException e) {
// thrown when the table to be altered does not exist
@@ -5426,7 +5424,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
private List<String> getPartValsFromName(RawStore ms, String catName, String dbName,
String tblName, String partName)
throws MetaException, InvalidObjectException {
- Table t = ms.getTable(catName, dbName, tblName, -1, null);
+ Table t = ms.getTable(catName, dbName, tblName, null);
if (t == null) {
throw new InvalidObjectException(dbName + "." + tblName
+ " table not found");
@@ -5684,7 +5682,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
try {
statsObj = getMS().getTableColumnStatistics(
parsedDbName[CAT_NAME], parsedDbName[DB_NAME], tableName, Lists.newArrayList(colName),
- -1, null);
+ null);
if (statsObj != null) {
assert statsObj.getStatsObjSize() <= 1;
}
@@ -5710,7 +5708,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
try {
ColumnStatistics cs = getMS().getTableColumnStatistics(
catName, dbName, tblName, lowerCaseColNames,
- request.getTxnId(), request.getValidWriteIdList());
+ request.getValidWriteIdList());
// Note: stats compliance is not propagated to the client; instead, we just return nothing
// if stats are not compliant for now. This won't work for stats merging, but that
// is currently only done on metastore size (see set_aggr...).
@@ -5779,7 +5777,6 @@ public class HiveMetaStore extends ThriftHiveMetastore {
try {
List<ColumnStatistics> stats = getMS().getPartitionColumnStatistics(
catName, dbName, tblName, lowerCasePartNames, lowerCaseColNames,
- request.isSetTxnId() ? request.getTxnId() : -1,
request.isSetValidWriteIdList() ? request.getValidWriteIdList() : null);
Map<String, List<ColumnStatisticsObj>> map = new HashMap<>();
if (stats != null) {
@@ -5803,7 +5800,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
@Override
public boolean update_table_column_statistics(ColumnStatistics colStats) throws TException {
// Deprecated API, won't work for transactional tables
- return updateTableColumnStatsInternal(colStats, -1, null, -1);
+ return updateTableColumnStatsInternal(colStats, null, -1);
}
@Override
@@ -5819,12 +5816,12 @@ public class HiveMetaStore extends ThriftHiveMetastore {
}
ColumnStatistics colStats = req.getColStatsIterator().next();
boolean ret = updateTableColumnStatsInternal(colStats,
- req.getTxnId(), req.getValidWriteIdList(), req.getWriteId());
+ req.getValidWriteIdList(), req.getWriteId());
return new SetPartitionsStatsResponse(ret);
}
private boolean updateTableColumnStatsInternal(ColumnStatistics colStats,
- long txnId, String validWriteIds, long writeId)
+ String validWriteIds, long writeId)
throws NoSuchObjectException, MetaException, InvalidObjectException, InvalidInputException {
normalizeColStatsInput(colStats);
@@ -5834,7 +5831,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
boolean ret = false;
try {
- ret = getMS().updateTableColumnStatistics(colStats, txnId, validWriteIds, writeId);
+ ret = getMS().updateTableColumnStatistics(colStats, validWriteIds, writeId);
} finally {
endFunction("write_column_statistics", ret != false, null,
colStats.getStatsDesc().getTableName());
@@ -5861,7 +5858,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
}
private boolean updatePartitonColStatsInternal(Table tbl, ColumnStatistics colStats,
- long txnId, String validWriteIds, long writeId)
+ String validWriteIds, long writeId)
throws MetaException, InvalidObjectException, NoSuchObjectException, InvalidInputException {
normalizeColStatsInput(colStats);
@@ -5878,7 +5875,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
}
List<String> partVals = getPartValsFromName(tbl, csd.getPartName());
return getMS().updatePartitionColumnStatistics(
- colStats, partVals, txnId, validWriteIds, writeId);
+ colStats, partVals, validWriteIds, writeId);
} finally {
endFunction("write_partition_column_statistics", ret != false, null, tableName);
}
@@ -5887,7 +5884,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
@Override
public boolean update_partition_column_statistics(ColumnStatistics colStats) throws TException {
// Deprecated API.
- return updatePartitonColStatsInternal(null, colStats, -1, null, -1);
+ return updatePartitonColStatsInternal(null, colStats, null, -1);
}
@@ -5904,7 +5901,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
}
ColumnStatistics colStats = req.getColStatsIterator().next();
boolean ret = updatePartitonColStatsInternal(null, colStats,
- req.getTxnId(), req.getValidWriteIdList(), req.getWriteId());
+ req.getValidWriteIdList(), req.getWriteId());
return new SetPartitionsStatsResponse(ret);
}
@@ -7560,8 +7557,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
try {
aggrStats = getMS().get_aggr_stats_for(catName, dbName, tblName,
- lowerCasePartNames, lowerCaseColNames, request.getTxnId(),
- request.getValidWriteIdList());
+ lowerCasePartNames, lowerCaseColNames, request.getValidWriteIdList());
return aggrStats;
} finally {
endFunction("get_aggr_stats_for", aggrStats == null, null, request.getTblName());
@@ -7597,12 +7593,11 @@ public class HiveMetaStore extends ThriftHiveMetastore {
return updateTableColumnStatsWithMerge(catName, dbName, tableName, colNames, request);
} else {
// This is the overwrite case, we do not care about the accuracy.
- return updateTableColumnStatsInternal(firstColStats, request.getTxnId(),
+ return updateTableColumnStatsInternal(firstColStats,
request.getValidWriteIdList(), request.getWriteId());
}
} else {
// partition level column stats merging
- List<Partition> partitions = new ArrayList<>();
// note that we may have two or more duplicate partition names.
// see autoColumnStats_2.q under TestMiniLlapLocalCliDriver
Map<String, ColumnStatistics> newStatsMap = new HashMap<>();
@@ -7621,7 +7616,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
Table t = getTable(catName, dbName, tableName);
for (Entry<String, ColumnStatistics> entry : newStatsMap.entrySet()) {
// We don't short-circuit on errors here anymore. That can leave acid stats invalid.
- ret = updatePartitonColStatsInternal(t, entry.getValue(), request.getTxnId(),
+ ret = updatePartitonColStatsInternal(t, entry.getValue(),
request.getValidWriteIdList(), request.getWriteId()) && ret;
}
}
@@ -7640,7 +7635,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
List<String> partitionNames = new ArrayList<>();
partitionNames.addAll(newStatsMap.keySet());
List<ColumnStatistics> csOlds = ms.getPartitionColumnStatistics(catName, dbName, tableName,
- partitionNames, colNames, request.getTxnId(), request.getValidWriteIdList());
+ partitionNames, colNames, request.getValidWriteIdList());
if (newStatsMap.values().size() != csOlds.size()) {
// some of the partitions miss stats.
LOG.debug("Some of the partitions miss stats.");
@@ -7678,7 +7673,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
if (!csNew.getStatsObj().isEmpty()) {
// We don't short-circuit on errors here anymore. That can leave acid stats invalid.
- result = updatePartitonColStatsInternal(t, csNew, request.getTxnId(),
+ result = updatePartitonColStatsInternal(t, csNew,
request.getValidWriteIdList(), request.getWriteId()) && result;
} else if (isInvalidTxnStats) {
// For now because the stats state is such as it is, we will invalidate everything.
@@ -7688,7 +7683,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
StatsSetupConst.clearColumnStatsState(part.getParameters());
StatsSetupConst.setBasicStatsState(part.getParameters(), StatsSetupConst.FALSE);
ms.alterPartition(catName, dbName, tableName, part.getValues(), part,
- request.getTxnId(), request.getValidWriteIdList());
+ request.getValidWriteIdList());
result = false;
} else {
// TODO: why doesn't the original call for non acid tables invalidate the stats?
@@ -7716,7 +7711,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
boolean isCommitted = false, result = false;
try {
ColumnStatistics csOld = ms.getTableColumnStatistics(catName, dbName, tableName, colNames,
- request.getTxnId(), request.getValidWriteIdList());
+ request.getValidWriteIdList());
// we first use the valid stats list to prune the stats
boolean isInvalidTxnStats = csOld != null
&& csOld.isSetIsStatsCompliant() && !csOld.isIsStatsCompliant();
@@ -7734,7 +7729,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
}
if (!firstColStats.getStatsObj().isEmpty()) {
- result = updateTableColumnStatsInternal(firstColStats, request.getTxnId(),
+ result = updateTableColumnStatsInternal(firstColStats,
request.getValidWriteIdList(), request.getWriteId());
} else if (isInvalidTxnStats) {
// For now because the stats state is such as it is, we will invalidate everything.
@@ -7744,7 +7739,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
t.setWriteId(request.getWriteId());
StatsSetupConst.clearColumnStatsState(t.getParameters());
StatsSetupConst.setBasicStatsState(t.getParameters(), StatsSetupConst.FALSE);
- ms.alterTable(catName, dbName, tableName, t, request.getTxnId(), request.getValidWriteIdList());
+ ms.alterTable(catName, dbName, tableName, t, request.getValidWriteIdList());
} else {
// TODO: why doesn't the original call for non acid tables invalidate the stats?
LOG.debug("All the column stats are not accurate to merge.");
@@ -7763,13 +7758,13 @@ public class HiveMetaStore extends ThriftHiveMetastore {
private Table getTable(String catName, String dbName, String tableName)
throws MetaException, InvalidObjectException {
- return getTable(catName, dbName, tableName, -1, null);
+ return getTable(catName, dbName, tableName, null);
}
private Table getTable(String catName, String dbName, String tableName,
- long txnId, String writeIdList)
+ String writeIdList)
throws MetaException, InvalidObjectException {
- Table t = getMS().getTable(catName, dbName, tableName, txnId, writeIdList);
+ Table t = getMS().getTable(catName, dbName, tableName, writeIdList);
if (t == null) {
throw new InvalidObjectException(TableName.getQualified(catName, dbName, tableName)
+ " table not found");