You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by se...@apache.org on 2018/07/25 18:27:41 UTC

[32/50] [abbrv] hive git commit: HIVE-20047 : remove txnID argument for txn stats methods (Sergey Shelukhin)

http://git-wip-us.apache.org/repos/asf/hive/blob/cdb32a7f/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ttypes.py
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ttypes.py b/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ttypes.py
index 923fe2c..7fc1e43 100644
--- a/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ttypes.py
+++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ttypes.py
@@ -2510,7 +2510,6 @@ class TruncateTableRequest:
    - dbName
    - tableName
    - partNames
-   - txnId
    - writeId
    - validWriteIdList
   """
@@ -2520,16 +2519,14 @@ class TruncateTableRequest:
     (1, TType.STRING, 'dbName', None, None, ), # 1
     (2, TType.STRING, 'tableName', None, None, ), # 2
     (3, TType.LIST, 'partNames', (TType.STRING,None), None, ), # 3
-    (4, TType.I64, 'txnId', None, -1, ), # 4
-    (5, TType.I64, 'writeId', None, -1, ), # 5
-    (6, TType.STRING, 'validWriteIdList', None, None, ), # 6
+    (4, TType.I64, 'writeId', None, -1, ), # 4
+    (5, TType.STRING, 'validWriteIdList', None, None, ), # 5
   )
 
-  def __init__(self, dbName=None, tableName=None, partNames=None, txnId=thrift_spec[4][4], writeId=thrift_spec[5][4], validWriteIdList=None,):
+  def __init__(self, dbName=None, tableName=None, partNames=None, writeId=thrift_spec[4][4], validWriteIdList=None,):
     self.dbName = dbName
     self.tableName = tableName
     self.partNames = partNames
-    self.txnId = txnId
     self.writeId = writeId
     self.validWriteIdList = validWriteIdList
 
@@ -2564,15 +2561,10 @@ class TruncateTableRequest:
           iprot.skip(ftype)
       elif fid == 4:
         if ftype == TType.I64:
-          self.txnId = iprot.readI64()
-        else:
-          iprot.skip(ftype)
-      elif fid == 5:
-        if ftype == TType.I64:
           self.writeId = iprot.readI64()
         else:
           iprot.skip(ftype)
-      elif fid == 6:
+      elif fid == 5:
         if ftype == TType.STRING:
           self.validWriteIdList = iprot.readString()
         else:
@@ -2602,16 +2594,12 @@ class TruncateTableRequest:
         oprot.writeString(iter75)
       oprot.writeListEnd()
       oprot.writeFieldEnd()
-    if self.txnId is not None:
-      oprot.writeFieldBegin('txnId', TType.I64, 4)
-      oprot.writeI64(self.txnId)
-      oprot.writeFieldEnd()
     if self.writeId is not None:
-      oprot.writeFieldBegin('writeId', TType.I64, 5)
+      oprot.writeFieldBegin('writeId', TType.I64, 4)
       oprot.writeI64(self.writeId)
       oprot.writeFieldEnd()
     if self.validWriteIdList is not None:
-      oprot.writeFieldBegin('validWriteIdList', TType.STRING, 6)
+      oprot.writeFieldBegin('validWriteIdList', TType.STRING, 5)
       oprot.writeString(self.validWriteIdList)
       oprot.writeFieldEnd()
     oprot.writeFieldStop()
@@ -2630,7 +2618,6 @@ class TruncateTableRequest:
     value = (value * 31) ^ hash(self.dbName)
     value = (value * 31) ^ hash(self.tableName)
     value = (value * 31) ^ hash(self.partNames)
-    value = (value * 31) ^ hash(self.txnId)
     value = (value * 31) ^ hash(self.writeId)
     value = (value * 31) ^ hash(self.validWriteIdList)
     return value
@@ -7316,7 +7303,6 @@ class SetPartitionsStatsRequest:
   Attributes:
    - colStats
    - needMerge
-   - txnId
    - writeId
    - validWriteIdList
   """
@@ -7325,15 +7311,13 @@ class SetPartitionsStatsRequest:
     None, # 0
     (1, TType.LIST, 'colStats', (TType.STRUCT,(ColumnStatistics, ColumnStatistics.thrift_spec)), None, ), # 1
     (2, TType.BOOL, 'needMerge', None, None, ), # 2
-    (3, TType.I64, 'txnId', None, -1, ), # 3
-    (4, TType.I64, 'writeId', None, -1, ), # 4
-    (5, TType.STRING, 'validWriteIdList', None, None, ), # 5
+    (3, TType.I64, 'writeId', None, -1, ), # 3
+    (4, TType.STRING, 'validWriteIdList', None, None, ), # 4
   )
 
-  def __init__(self, colStats=None, needMerge=None, txnId=thrift_spec[3][4], writeId=thrift_spec[4][4], validWriteIdList=None,):
+  def __init__(self, colStats=None, needMerge=None, writeId=thrift_spec[3][4], validWriteIdList=None,):
     self.colStats = colStats
     self.needMerge = needMerge
-    self.txnId = txnId
     self.writeId = writeId
     self.validWriteIdList = validWriteIdList
 
@@ -7364,15 +7348,10 @@ class SetPartitionsStatsRequest:
           iprot.skip(ftype)
       elif fid == 3:
         if ftype == TType.I64:
-          self.txnId = iprot.readI64()
-        else:
-          iprot.skip(ftype)
-      elif fid == 4:
-        if ftype == TType.I64:
           self.writeId = iprot.readI64()
         else:
           iprot.skip(ftype)
-      elif fid == 5:
+      elif fid == 4:
         if ftype == TType.STRING:
           self.validWriteIdList = iprot.readString()
         else:
@@ -7398,16 +7377,12 @@ class SetPartitionsStatsRequest:
       oprot.writeFieldBegin('needMerge', TType.BOOL, 2)
       oprot.writeBool(self.needMerge)
       oprot.writeFieldEnd()
-    if self.txnId is not None:
-      oprot.writeFieldBegin('txnId', TType.I64, 3)
-      oprot.writeI64(self.txnId)
-      oprot.writeFieldEnd()
     if self.writeId is not None:
-      oprot.writeFieldBegin('writeId', TType.I64, 4)
+      oprot.writeFieldBegin('writeId', TType.I64, 3)
       oprot.writeI64(self.writeId)
       oprot.writeFieldEnd()
     if self.validWriteIdList is not None:
-      oprot.writeFieldBegin('validWriteIdList', TType.STRING, 5)
+      oprot.writeFieldBegin('validWriteIdList', TType.STRING, 4)
       oprot.writeString(self.validWriteIdList)
       oprot.writeFieldEnd()
     oprot.writeFieldStop()
@@ -7423,7 +7398,6 @@ class SetPartitionsStatsRequest:
     value = 17
     value = (value * 31) ^ hash(self.colStats)
     value = (value * 31) ^ hash(self.needMerge)
-    value = (value * 31) ^ hash(self.txnId)
     value = (value * 31) ^ hash(self.writeId)
     value = (value * 31) ^ hash(self.validWriteIdList)
     return value
@@ -9722,7 +9696,6 @@ class TableStatsRequest:
    - tblName
    - colNames
    - catName
-   - txnId
    - validWriteIdList
   """
 
@@ -9732,16 +9705,14 @@ class TableStatsRequest:
     (2, TType.STRING, 'tblName', None, None, ), # 2
     (3, TType.LIST, 'colNames', (TType.STRING,None), None, ), # 3
     (4, TType.STRING, 'catName', None, None, ), # 4
-    (5, TType.I64, 'txnId', None, -1, ), # 5
-    (6, TType.STRING, 'validWriteIdList', None, None, ), # 6
+    (5, TType.STRING, 'validWriteIdList', None, None, ), # 5
   )
 
-  def __init__(self, dbName=None, tblName=None, colNames=None, catName=None, txnId=thrift_spec[5][4], validWriteIdList=None,):
+  def __init__(self, dbName=None, tblName=None, colNames=None, catName=None, validWriteIdList=None,):
     self.dbName = dbName
     self.tblName = tblName
     self.colNames = colNames
     self.catName = catName
-    self.txnId = txnId
     self.validWriteIdList = validWriteIdList
 
   def read(self, iprot):
@@ -9779,11 +9750,6 @@ class TableStatsRequest:
         else:
           iprot.skip(ftype)
       elif fid == 5:
-        if ftype == TType.I64:
-          self.txnId = iprot.readI64()
-        else:
-          iprot.skip(ftype)
-      elif fid == 6:
         if ftype == TType.STRING:
           self.validWriteIdList = iprot.readString()
         else:
@@ -9817,12 +9783,8 @@ class TableStatsRequest:
       oprot.writeFieldBegin('catName', TType.STRING, 4)
       oprot.writeString(self.catName)
       oprot.writeFieldEnd()
-    if self.txnId is not None:
-      oprot.writeFieldBegin('txnId', TType.I64, 5)
-      oprot.writeI64(self.txnId)
-      oprot.writeFieldEnd()
     if self.validWriteIdList is not None:
-      oprot.writeFieldBegin('validWriteIdList', TType.STRING, 6)
+      oprot.writeFieldBegin('validWriteIdList', TType.STRING, 5)
       oprot.writeString(self.validWriteIdList)
       oprot.writeFieldEnd()
     oprot.writeFieldStop()
@@ -9844,7 +9806,6 @@ class TableStatsRequest:
     value = (value * 31) ^ hash(self.tblName)
     value = (value * 31) ^ hash(self.colNames)
     value = (value * 31) ^ hash(self.catName)
-    value = (value * 31) ^ hash(self.txnId)
     value = (value * 31) ^ hash(self.validWriteIdList)
     return value
 
@@ -9867,7 +9828,6 @@ class PartitionsStatsRequest:
    - colNames
    - partNames
    - catName
-   - txnId
    - validWriteIdList
   """
 
@@ -9878,17 +9838,15 @@ class PartitionsStatsRequest:
     (3, TType.LIST, 'colNames', (TType.STRING,None), None, ), # 3
     (4, TType.LIST, 'partNames', (TType.STRING,None), None, ), # 4
     (5, TType.STRING, 'catName', None, None, ), # 5
-    (6, TType.I64, 'txnId', None, -1, ), # 6
-    (7, TType.STRING, 'validWriteIdList', None, None, ), # 7
+    (6, TType.STRING, 'validWriteIdList', None, None, ), # 6
   )
 
-  def __init__(self, dbName=None, tblName=None, colNames=None, partNames=None, catName=None, txnId=thrift_spec[6][4], validWriteIdList=None,):
+  def __init__(self, dbName=None, tblName=None, colNames=None, partNames=None, catName=None, validWriteIdList=None,):
     self.dbName = dbName
     self.tblName = tblName
     self.colNames = colNames
     self.partNames = partNames
     self.catName = catName
-    self.txnId = txnId
     self.validWriteIdList = validWriteIdList
 
   def read(self, iprot):
@@ -9936,11 +9894,6 @@ class PartitionsStatsRequest:
         else:
           iprot.skip(ftype)
       elif fid == 6:
-        if ftype == TType.I64:
-          self.txnId = iprot.readI64()
-        else:
-          iprot.skip(ftype)
-      elif fid == 7:
         if ftype == TType.STRING:
           self.validWriteIdList = iprot.readString()
         else:
@@ -9981,12 +9934,8 @@ class PartitionsStatsRequest:
       oprot.writeFieldBegin('catName', TType.STRING, 5)
       oprot.writeString(self.catName)
       oprot.writeFieldEnd()
-    if self.txnId is not None:
-      oprot.writeFieldBegin('txnId', TType.I64, 6)
-      oprot.writeI64(self.txnId)
-      oprot.writeFieldEnd()
     if self.validWriteIdList is not None:
-      oprot.writeFieldBegin('validWriteIdList', TType.STRING, 7)
+      oprot.writeFieldBegin('validWriteIdList', TType.STRING, 6)
       oprot.writeString(self.validWriteIdList)
       oprot.writeFieldEnd()
     oprot.writeFieldStop()
@@ -10011,7 +9960,6 @@ class PartitionsStatsRequest:
     value = (value * 31) ^ hash(self.colNames)
     value = (value * 31) ^ hash(self.partNames)
     value = (value * 31) ^ hash(self.catName)
-    value = (value * 31) ^ hash(self.txnId)
     value = (value * 31) ^ hash(self.validWriteIdList)
     return value
 
@@ -10122,7 +10070,6 @@ class AddPartitionsRequest:
    - ifNotExists
    - needResult
    - catName
-   - txnId
    - validWriteIdList
   """
 
@@ -10134,18 +10081,16 @@ class AddPartitionsRequest:
     (4, TType.BOOL, 'ifNotExists', None, None, ), # 4
     (5, TType.BOOL, 'needResult', None, True, ), # 5
     (6, TType.STRING, 'catName', None, None, ), # 6
-    (7, TType.I64, 'txnId', None, -1, ), # 7
-    (8, TType.STRING, 'validWriteIdList', None, None, ), # 8
+    (7, TType.STRING, 'validWriteIdList', None, None, ), # 7
   )
 
-  def __init__(self, dbName=None, tblName=None, parts=None, ifNotExists=None, needResult=thrift_spec[5][4], catName=None, txnId=thrift_spec[7][4], validWriteIdList=None,):
+  def __init__(self, dbName=None, tblName=None, parts=None, ifNotExists=None, needResult=thrift_spec[5][4], catName=None, validWriteIdList=None,):
     self.dbName = dbName
     self.tblName = tblName
     self.parts = parts
     self.ifNotExists = ifNotExists
     self.needResult = needResult
     self.catName = catName
-    self.txnId = txnId
     self.validWriteIdList = validWriteIdList
 
   def read(self, iprot):
@@ -10194,11 +10139,6 @@ class AddPartitionsRequest:
         else:
           iprot.skip(ftype)
       elif fid == 7:
-        if ftype == TType.I64:
-          self.txnId = iprot.readI64()
-        else:
-          iprot.skip(ftype)
-      elif fid == 8:
         if ftype == TType.STRING:
           self.validWriteIdList = iprot.readString()
         else:
@@ -10240,12 +10180,8 @@ class AddPartitionsRequest:
       oprot.writeFieldBegin('catName', TType.STRING, 6)
       oprot.writeString(self.catName)
       oprot.writeFieldEnd()
-    if self.txnId is not None:
-      oprot.writeFieldBegin('txnId', TType.I64, 7)
-      oprot.writeI64(self.txnId)
-      oprot.writeFieldEnd()
     if self.validWriteIdList is not None:
-      oprot.writeFieldBegin('validWriteIdList', TType.STRING, 8)
+      oprot.writeFieldBegin('validWriteIdList', TType.STRING, 7)
       oprot.writeString(self.validWriteIdList)
       oprot.writeFieldEnd()
     oprot.writeFieldStop()
@@ -10271,7 +10207,6 @@ class AddPartitionsRequest:
     value = (value * 31) ^ hash(self.ifNotExists)
     value = (value * 31) ^ hash(self.needResult)
     value = (value * 31) ^ hash(self.catName)
-    value = (value * 31) ^ hash(self.txnId)
     value = (value * 31) ^ hash(self.validWriteIdList)
     return value
 
@@ -17526,7 +17461,6 @@ class GetTableRequest:
    - tblName
    - capabilities
    - catName
-   - txnId
    - validWriteIdList
   """
 
@@ -17536,16 +17470,15 @@ class GetTableRequest:
     (2, TType.STRING, 'tblName', None, None, ), # 2
     (3, TType.STRUCT, 'capabilities', (ClientCapabilities, ClientCapabilities.thrift_spec), None, ), # 3
     (4, TType.STRING, 'catName', None, None, ), # 4
-    (5, TType.I64, 'txnId', None, -1, ), # 5
+    None, # 5
     (6, TType.STRING, 'validWriteIdList', None, None, ), # 6
   )
 
-  def __init__(self, dbName=None, tblName=None, capabilities=None, catName=None, txnId=thrift_spec[5][4], validWriteIdList=None,):
+  def __init__(self, dbName=None, tblName=None, capabilities=None, catName=None, validWriteIdList=None,):
     self.dbName = dbName
     self.tblName = tblName
     self.capabilities = capabilities
     self.catName = catName
-    self.txnId = txnId
     self.validWriteIdList = validWriteIdList
 
   def read(self, iprot):
@@ -17578,11 +17511,6 @@ class GetTableRequest:
           self.catName = iprot.readString()
         else:
           iprot.skip(ftype)
-      elif fid == 5:
-        if ftype == TType.I64:
-          self.txnId = iprot.readI64()
-        else:
-          iprot.skip(ftype)
       elif fid == 6:
         if ftype == TType.STRING:
           self.validWriteIdList = iprot.readString()
@@ -17614,10 +17542,6 @@ class GetTableRequest:
       oprot.writeFieldBegin('catName', TType.STRING, 4)
       oprot.writeString(self.catName)
       oprot.writeFieldEnd()
-    if self.txnId is not None:
-      oprot.writeFieldBegin('txnId', TType.I64, 5)
-      oprot.writeI64(self.txnId)
-      oprot.writeFieldEnd()
     if self.validWriteIdList is not None:
       oprot.writeFieldBegin('validWriteIdList', TType.STRING, 6)
       oprot.writeString(self.validWriteIdList)
@@ -17639,7 +17563,6 @@ class GetTableRequest:
     value = (value * 31) ^ hash(self.tblName)
     value = (value * 31) ^ hash(self.capabilities)
     value = (value * 31) ^ hash(self.catName)
-    value = (value * 31) ^ hash(self.txnId)
     value = (value * 31) ^ hash(self.validWriteIdList)
     return value
 
@@ -22583,7 +22506,6 @@ class AlterPartitionsRequest:
    - tableName
    - partitions
    - environmentContext
-   - txnId
    - writeId
    - validWriteIdList
   """
@@ -22595,18 +22517,16 @@ class AlterPartitionsRequest:
     (3, TType.STRING, 'tableName', None, None, ), # 3
     (4, TType.LIST, 'partitions', (TType.STRUCT,(Partition, Partition.thrift_spec)), None, ), # 4
     (5, TType.STRUCT, 'environmentContext', (EnvironmentContext, EnvironmentContext.thrift_spec), None, ), # 5
-    (6, TType.I64, 'txnId', None, -1, ), # 6
-    (7, TType.I64, 'writeId', None, -1, ), # 7
-    (8, TType.STRING, 'validWriteIdList', None, None, ), # 8
+    (6, TType.I64, 'writeId', None, -1, ), # 6
+    (7, TType.STRING, 'validWriteIdList', None, None, ), # 7
   )
 
-  def __init__(self, catName=None, dbName=None, tableName=None, partitions=None, environmentContext=None, txnId=thrift_spec[6][4], writeId=thrift_spec[7][4], validWriteIdList=None,):
+  def __init__(self, catName=None, dbName=None, tableName=None, partitions=None, environmentContext=None, writeId=thrift_spec[6][4], validWriteIdList=None,):
     self.catName = catName
     self.dbName = dbName
     self.tableName = tableName
     self.partitions = partitions
     self.environmentContext = environmentContext
-    self.txnId = txnId
     self.writeId = writeId
     self.validWriteIdList = validWriteIdList
 
@@ -22653,15 +22573,10 @@ class AlterPartitionsRequest:
           iprot.skip(ftype)
       elif fid == 6:
         if ftype == TType.I64:
-          self.txnId = iprot.readI64()
-        else:
-          iprot.skip(ftype)
-      elif fid == 7:
-        if ftype == TType.I64:
           self.writeId = iprot.readI64()
         else:
           iprot.skip(ftype)
-      elif fid == 8:
+      elif fid == 7:
         if ftype == TType.STRING:
           self.validWriteIdList = iprot.readString()
         else:
@@ -22699,16 +22614,12 @@ class AlterPartitionsRequest:
       oprot.writeFieldBegin('environmentContext', TType.STRUCT, 5)
       self.environmentContext.write(oprot)
       oprot.writeFieldEnd()
-    if self.txnId is not None:
-      oprot.writeFieldBegin('txnId', TType.I64, 6)
-      oprot.writeI64(self.txnId)
-      oprot.writeFieldEnd()
     if self.writeId is not None:
-      oprot.writeFieldBegin('writeId', TType.I64, 7)
+      oprot.writeFieldBegin('writeId', TType.I64, 6)
       oprot.writeI64(self.writeId)
       oprot.writeFieldEnd()
     if self.validWriteIdList is not None:
-      oprot.writeFieldBegin('validWriteIdList', TType.STRING, 8)
+      oprot.writeFieldBegin('validWriteIdList', TType.STRING, 7)
       oprot.writeString(self.validWriteIdList)
       oprot.writeFieldEnd()
     oprot.writeFieldStop()
@@ -22731,7 +22642,6 @@ class AlterPartitionsRequest:
     value = (value * 31) ^ hash(self.tableName)
     value = (value * 31) ^ hash(self.partitions)
     value = (value * 31) ^ hash(self.environmentContext)
-    value = (value * 31) ^ hash(self.txnId)
     value = (value * 31) ^ hash(self.writeId)
     value = (value * 31) ^ hash(self.validWriteIdList)
     return value
@@ -22801,7 +22711,6 @@ class RenamePartitionRequest:
    - tableName
    - partVals
    - newPart
-   - txnId
    - validWriteIdList
   """
 
@@ -22812,17 +22721,15 @@ class RenamePartitionRequest:
     (3, TType.STRING, 'tableName', None, None, ), # 3
     (4, TType.LIST, 'partVals', (TType.STRING,None), None, ), # 4
     (5, TType.STRUCT, 'newPart', (Partition, Partition.thrift_spec), None, ), # 5
-    (6, TType.I64, 'txnId', None, -1, ), # 6
-    (7, TType.STRING, 'validWriteIdList', None, None, ), # 7
+    (6, TType.STRING, 'validWriteIdList', None, None, ), # 6
   )
 
-  def __init__(self, catName=None, dbName=None, tableName=None, partVals=None, newPart=None, txnId=thrift_spec[6][4], validWriteIdList=None,):
+  def __init__(self, catName=None, dbName=None, tableName=None, partVals=None, newPart=None, validWriteIdList=None,):
     self.catName = catName
     self.dbName = dbName
     self.tableName = tableName
     self.partVals = partVals
     self.newPart = newPart
-    self.txnId = txnId
     self.validWriteIdList = validWriteIdList
 
   def read(self, iprot):
@@ -22866,11 +22773,6 @@ class RenamePartitionRequest:
         else:
           iprot.skip(ftype)
       elif fid == 6:
-        if ftype == TType.I64:
-          self.txnId = iprot.readI64()
-        else:
-          iprot.skip(ftype)
-      elif fid == 7:
         if ftype == TType.STRING:
           self.validWriteIdList = iprot.readString()
         else:
@@ -22908,12 +22810,8 @@ class RenamePartitionRequest:
       oprot.writeFieldBegin('newPart', TType.STRUCT, 5)
       self.newPart.write(oprot)
       oprot.writeFieldEnd()
-    if self.txnId is not None:
-      oprot.writeFieldBegin('txnId', TType.I64, 6)
-      oprot.writeI64(self.txnId)
-      oprot.writeFieldEnd()
     if self.validWriteIdList is not None:
-      oprot.writeFieldBegin('validWriteIdList', TType.STRING, 7)
+      oprot.writeFieldBegin('validWriteIdList', TType.STRING, 6)
       oprot.writeString(self.validWriteIdList)
       oprot.writeFieldEnd()
     oprot.writeFieldStop()
@@ -22938,7 +22836,6 @@ class RenamePartitionRequest:
     value = (value * 31) ^ hash(self.tableName)
     value = (value * 31) ^ hash(self.partVals)
     value = (value * 31) ^ hash(self.newPart)
-    value = (value * 31) ^ hash(self.txnId)
     value = (value * 31) ^ hash(self.validWriteIdList)
     return value
 
@@ -23007,7 +22904,6 @@ class AlterTableRequest:
    - tableName
    - table
    - environmentContext
-   - txnId
    - writeId
    - validWriteIdList
   """
@@ -23019,18 +22915,16 @@ class AlterTableRequest:
     (3, TType.STRING, 'tableName', None, None, ), # 3
     (4, TType.STRUCT, 'table', (Table, Table.thrift_spec), None, ), # 4
     (5, TType.STRUCT, 'environmentContext', (EnvironmentContext, EnvironmentContext.thrift_spec), None, ), # 5
-    (6, TType.I64, 'txnId', None, -1, ), # 6
-    (7, TType.I64, 'writeId', None, -1, ), # 7
-    (8, TType.STRING, 'validWriteIdList', None, None, ), # 8
+    (6, TType.I64, 'writeId', None, -1, ), # 6
+    (7, TType.STRING, 'validWriteIdList', None, None, ), # 7
   )
 
-  def __init__(self, catName=None, dbName=None, tableName=None, table=None, environmentContext=None, txnId=thrift_spec[6][4], writeId=thrift_spec[7][4], validWriteIdList=None,):
+  def __init__(self, catName=None, dbName=None, tableName=None, table=None, environmentContext=None, writeId=thrift_spec[6][4], validWriteIdList=None,):
     self.catName = catName
     self.dbName = dbName
     self.tableName = tableName
     self.table = table
     self.environmentContext = environmentContext
-    self.txnId = txnId
     self.writeId = writeId
     self.validWriteIdList = validWriteIdList
 
@@ -23072,15 +22966,10 @@ class AlterTableRequest:
           iprot.skip(ftype)
       elif fid == 6:
         if ftype == TType.I64:
-          self.txnId = iprot.readI64()
-        else:
-          iprot.skip(ftype)
-      elif fid == 7:
-        if ftype == TType.I64:
           self.writeId = iprot.readI64()
         else:
           iprot.skip(ftype)
-      elif fid == 8:
+      elif fid == 7:
         if ftype == TType.STRING:
           self.validWriteIdList = iprot.readString()
         else:
@@ -23115,16 +23004,12 @@ class AlterTableRequest:
       oprot.writeFieldBegin('environmentContext', TType.STRUCT, 5)
       self.environmentContext.write(oprot)
       oprot.writeFieldEnd()
-    if self.txnId is not None:
-      oprot.writeFieldBegin('txnId', TType.I64, 6)
-      oprot.writeI64(self.txnId)
-      oprot.writeFieldEnd()
     if self.writeId is not None:
-      oprot.writeFieldBegin('writeId', TType.I64, 7)
+      oprot.writeFieldBegin('writeId', TType.I64, 6)
       oprot.writeI64(self.writeId)
       oprot.writeFieldEnd()
     if self.validWriteIdList is not None:
-      oprot.writeFieldBegin('validWriteIdList', TType.STRING, 8)
+      oprot.writeFieldBegin('validWriteIdList', TType.STRING, 7)
       oprot.writeString(self.validWriteIdList)
       oprot.writeFieldEnd()
     oprot.writeFieldStop()
@@ -23147,7 +23032,6 @@ class AlterTableRequest:
     value = (value * 31) ^ hash(self.tableName)
     value = (value * 31) ^ hash(self.table)
     value = (value * 31) ^ hash(self.environmentContext)
-    value = (value * 31) ^ hash(self.txnId)
     value = (value * 31) ^ hash(self.writeId)
     value = (value * 31) ^ hash(self.validWriteIdList)
     return value

http://git-wip-us.apache.org/repos/asf/hive/blob/cdb32a7f/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/hive_metastore_types.rb
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/hive_metastore_types.rb b/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/hive_metastore_types.rb
index 2e2392d..e0c6c02 100644
--- a/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/hive_metastore_types.rb
+++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/hive_metastore_types.rb
@@ -608,15 +608,13 @@ class TruncateTableRequest
   DBNAME = 1
   TABLENAME = 2
   PARTNAMES = 3
-  TXNID = 4
-  WRITEID = 5
-  VALIDWRITEIDLIST = 6
+  WRITEID = 4
+  VALIDWRITEIDLIST = 5
 
   FIELDS = {
     DBNAME => {:type => ::Thrift::Types::STRING, :name => 'dbName'},
     TABLENAME => {:type => ::Thrift::Types::STRING, :name => 'tableName'},
     PARTNAMES => {:type => ::Thrift::Types::LIST, :name => 'partNames', :element => {:type => ::Thrift::Types::STRING}, :optional => true},
-    TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true},
     WRITEID => {:type => ::Thrift::Types::I64, :name => 'writeId', :default => -1, :optional => true},
     VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true}
   }
@@ -1646,14 +1644,12 @@ class SetPartitionsStatsRequest
   include ::Thrift::Struct, ::Thrift::Struct_Union
   COLSTATS = 1
   NEEDMERGE = 2
-  TXNID = 3
-  WRITEID = 4
-  VALIDWRITEIDLIST = 5
+  WRITEID = 3
+  VALIDWRITEIDLIST = 4
 
   FIELDS = {
     COLSTATS => {:type => ::Thrift::Types::LIST, :name => 'colStats', :element => {:type => ::Thrift::Types::STRUCT, :class => ::ColumnStatistics}},
     NEEDMERGE => {:type => ::Thrift::Types::BOOL, :name => 'needMerge', :optional => true},
-    TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true},
     WRITEID => {:type => ::Thrift::Types::I64, :name => 'writeId', :default => -1, :optional => true},
     VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true}
   }
@@ -2178,15 +2174,13 @@ class TableStatsRequest
   TBLNAME = 2
   COLNAMES = 3
   CATNAME = 4
-  TXNID = 5
-  VALIDWRITEIDLIST = 6
+  VALIDWRITEIDLIST = 5
 
   FIELDS = {
     DBNAME => {:type => ::Thrift::Types::STRING, :name => 'dbName'},
     TBLNAME => {:type => ::Thrift::Types::STRING, :name => 'tblName'},
     COLNAMES => {:type => ::Thrift::Types::LIST, :name => 'colNames', :element => {:type => ::Thrift::Types::STRING}},
     CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName', :optional => true},
-    TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true},
     VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true}
   }
 
@@ -2208,8 +2202,7 @@ class PartitionsStatsRequest
   COLNAMES = 3
   PARTNAMES = 4
   CATNAME = 5
-  TXNID = 6
-  VALIDWRITEIDLIST = 7
+  VALIDWRITEIDLIST = 6
 
   FIELDS = {
     DBNAME => {:type => ::Thrift::Types::STRING, :name => 'dbName'},
@@ -2217,7 +2210,6 @@ class PartitionsStatsRequest
     COLNAMES => {:type => ::Thrift::Types::LIST, :name => 'colNames', :element => {:type => ::Thrift::Types::STRING}},
     PARTNAMES => {:type => ::Thrift::Types::LIST, :name => 'partNames', :element => {:type => ::Thrift::Types::STRING}},
     CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName', :optional => true},
-    TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true},
     VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true}
   }
 
@@ -2259,8 +2251,7 @@ class AddPartitionsRequest
   IFNOTEXISTS = 4
   NEEDRESULT = 5
   CATNAME = 6
-  TXNID = 7
-  VALIDWRITEIDLIST = 8
+  VALIDWRITEIDLIST = 7
 
   FIELDS = {
     DBNAME => {:type => ::Thrift::Types::STRING, :name => 'dbName'},
@@ -2269,7 +2260,6 @@ class AddPartitionsRequest
     IFNOTEXISTS => {:type => ::Thrift::Types::BOOL, :name => 'ifNotExists'},
     NEEDRESULT => {:type => ::Thrift::Types::BOOL, :name => 'needResult', :default => true, :optional => true},
     CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName', :optional => true},
-    TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true},
     VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true}
   }
 
@@ -3915,7 +3905,6 @@ class GetTableRequest
   TBLNAME = 2
   CAPABILITIES = 3
   CATNAME = 4
-  TXNID = 5
   VALIDWRITEIDLIST = 6
 
   FIELDS = {
@@ -3923,7 +3912,6 @@ class GetTableRequest
     TBLNAME => {:type => ::Thrift::Types::STRING, :name => 'tblName'},
     CAPABILITIES => {:type => ::Thrift::Types::STRUCT, :name => 'capabilities', :class => ::ClientCapabilities, :optional => true},
     CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName', :optional => true},
-    TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true},
     VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true}
   }
 
@@ -5114,9 +5102,8 @@ class AlterPartitionsRequest
   TABLENAME = 3
   PARTITIONS = 4
   ENVIRONMENTCONTEXT = 5
-  TXNID = 6
-  WRITEID = 7
-  VALIDWRITEIDLIST = 8
+  WRITEID = 6
+  VALIDWRITEIDLIST = 7
 
   FIELDS = {
     CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName', :optional => true},
@@ -5124,7 +5111,6 @@ class AlterPartitionsRequest
     TABLENAME => {:type => ::Thrift::Types::STRING, :name => 'tableName'},
     PARTITIONS => {:type => ::Thrift::Types::LIST, :name => 'partitions', :element => {:type => ::Thrift::Types::STRUCT, :class => ::Partition}},
     ENVIRONMENTCONTEXT => {:type => ::Thrift::Types::STRUCT, :name => 'environmentContext', :class => ::EnvironmentContext, :optional => true},
-    TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true},
     WRITEID => {:type => ::Thrift::Types::I64, :name => 'writeId', :default => -1, :optional => true},
     VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true}
   }
@@ -5162,8 +5148,7 @@ class RenamePartitionRequest
   TABLENAME = 3
   PARTVALS = 4
   NEWPART = 5
-  TXNID = 6
-  VALIDWRITEIDLIST = 7
+  VALIDWRITEIDLIST = 6
 
   FIELDS = {
     CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName', :optional => true},
@@ -5171,7 +5156,6 @@ class RenamePartitionRequest
     TABLENAME => {:type => ::Thrift::Types::STRING, :name => 'tableName'},
     PARTVALS => {:type => ::Thrift::Types::LIST, :name => 'partVals', :element => {:type => ::Thrift::Types::STRING}},
     NEWPART => {:type => ::Thrift::Types::STRUCT, :name => 'newPart', :class => ::Partition},
-    TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true},
     VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true}
   }
 
@@ -5209,9 +5193,8 @@ class AlterTableRequest
   TABLENAME = 3
   TABLE = 4
   ENVIRONMENTCONTEXT = 5
-  TXNID = 6
-  WRITEID = 7
-  VALIDWRITEIDLIST = 8
+  WRITEID = 6
+  VALIDWRITEIDLIST = 7
 
   FIELDS = {
     CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName', :optional => true},
@@ -5219,7 +5202,6 @@ class AlterTableRequest
     TABLENAME => {:type => ::Thrift::Types::STRING, :name => 'tableName'},
     TABLE => {:type => ::Thrift::Types::STRUCT, :name => 'table', :class => ::Table},
     ENVIRONMENTCONTEXT => {:type => ::Thrift::Types::STRUCT, :name => 'environmentContext', :class => ::EnvironmentContext, :optional => true},
-    TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true},
     WRITEID => {:type => ::Thrift::Types::I64, :name => 'writeId', :default => -1, :optional => true},
     VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true}
   }

http://git-wip-us.apache.org/repos/asf/hive/blob/cdb32a7f/standalone-metastore/metastore-common/src/main/thrift/hive_metastore.thrift
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-common/src/main/thrift/hive_metastore.thrift b/standalone-metastore/metastore-common/src/main/thrift/hive_metastore.thrift
index 4cfa6a1..8a4bdd8 100644
--- a/standalone-metastore/metastore-common/src/main/thrift/hive_metastore.thrift
+++ b/standalone-metastore/metastore-common/src/main/thrift/hive_metastore.thrift
@@ -282,9 +282,8 @@ struct TruncateTableRequest {
   1: required string dbName,
   2: required string tableName,
   3: optional list<string> partNames,
-  4: optional i64 txnId=-1,
-  5: optional i64 writeId=-1,
-  6: optional string validWriteIdList
+  4: optional i64 writeId=-1,
+  5: optional string validWriteIdList
 }
 
 struct TruncateTableResponse {
@@ -596,9 +595,8 @@ struct AggrStats {
 struct SetPartitionsStatsRequest {
 1: required list<ColumnStatistics> colStats,
 2: optional bool needMerge, //stats need to be merged with the existing stats
-3: optional i64 txnId=-1,   // transaction id of the query that sends this structure
-4: optional i64 writeId=-1,         // writeId for the current query that updates the stats
-5: optional string validWriteIdList // valid write id list for the table for which this struct is being sent
+3: optional i64 writeId=-1,         // writeId for the current query that updates the stats
+4: optional string validWriteIdList // valid write id list for the table for which this struct is being sent
 }
 
 struct SetPartitionsStatsResponse {
@@ -745,8 +743,7 @@ struct TableStatsRequest {
  2: required string tblName,
  3: required list<string> colNames
  4: optional string catName,
- 5: optional i64 txnId=-1,            // transaction id of the query that sends this structure
- 6: optional string validWriteIdList  // valid write id list for the table for which this struct is being sent
+ 5: optional string validWriteIdList  // valid write id list for the table for which this struct is being sent
 }
 
 struct PartitionsStatsRequest {
@@ -755,8 +752,7 @@ struct PartitionsStatsRequest {
  3: required list<string> colNames,
  4: required list<string> partNames,
  5: optional string catName,
- 6: optional i64 txnId=-1,           // transaction id of the query that sends this structure
- 7: optional string validWriteIdList // valid write id list for the table for which this struct is being sent
+ 6: optional string validWriteIdList // valid write id list for the table for which this struct is being sent
 }
 
 // Return type for add_partitions_req
@@ -773,8 +769,7 @@ struct AddPartitionsRequest {
   4: required bool ifNotExists,
   5: optional bool needResult=true,
   6: optional string catName,
-  7: optional i64 txnId=-1,
-  8: optional string validWriteIdList
+  7: optional string validWriteIdList
 }
 
 // Return type for drop_partitions_req
@@ -1275,7 +1270,6 @@ struct GetTableRequest {
   2: required string tblName,
   3: optional ClientCapabilities capabilities,
   4: optional string catName,
-  5: optional i64 txnId=-1,
   6: optional string validWriteIdList
 }
 
@@ -1615,9 +1609,8 @@ struct AlterPartitionsRequest {
   3: required string tableName,
   4: required list<Partition> partitions,
   5: optional EnvironmentContext environmentContext,
-  6: optional i64 txnId=-1,
-  7: optional i64 writeId=-1,
-  8: optional string validWriteIdList
+  6: optional i64 writeId=-1,
+  7: optional string validWriteIdList
 }
 
 struct AlterPartitionsResponse {
@@ -1629,8 +1622,7 @@ struct RenamePartitionRequest {
   3: required string tableName,
   4: required list<string> partVals,
   5: required Partition newPart,
-  6: optional i64 txnId=-1,
-  7: optional string validWriteIdList
+  6: optional string validWriteIdList
 }
 
 struct RenamePartitionResponse {
@@ -1642,9 +1634,8 @@ struct AlterTableRequest {
   3: required string tableName,
   4: required Table table,
   5: optional EnvironmentContext environmentContext,
-  6: optional i64 txnId=-1,
-  7: optional i64 writeId=-1,
-  8: optional string validWriteIdList
+  6: optional i64 writeId=-1,
+  7: optional string validWriteIdList
 // TODO: also add cascade here, out of envCtx
 }
 

http://git-wip-us.apache.org/repos/asf/hive/blob/cdb32a7f/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/AlterHandler.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/AlterHandler.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/AlterHandler.java
index f3dc264..f4e0c41 100644
--- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/AlterHandler.java
+++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/AlterHandler.java
@@ -61,7 +61,7 @@ public interface AlterHandler extends Configurable {
   default void alterTable(RawStore msdb, Warehouse wh, String catName, String dbname,
     String name, Table newTable, EnvironmentContext envContext)
       throws InvalidOperationException, MetaException {
-    alterTable(msdb, wh, catName, dbname, name, newTable, envContext, null, -1, null);
+    alterTable(msdb, wh, catName, dbname, name, newTable, envContext, null, null);
   }
 
   /**
@@ -88,7 +88,7 @@ public interface AlterHandler extends Configurable {
    */
   void alterTable(RawStore msdb, Warehouse wh, String catName, String dbname,
       String name, Table newTable, EnvironmentContext envContext,
-      IHMSHandler handler, long txnId, String writeIdList)
+      IHMSHandler handler,  String writeIdList)
           throws InvalidOperationException, MetaException;
 
   /**
@@ -146,7 +146,7 @@ public interface AlterHandler extends Configurable {
   Partition alterPartition(final RawStore msdb, Warehouse wh, final String catName,
                            final String dbname, final String name, final List<String> part_vals,
                            final Partition new_part, EnvironmentContext environmentContext,
-                           IHMSHandler handler, long txnId, String validWriteIds)
+                           IHMSHandler handler,  String validWriteIds)
       throws InvalidOperationException, InvalidObjectException, AlreadyExistsException, MetaException;
 
   /**
@@ -198,7 +198,7 @@ public interface AlterHandler extends Configurable {
    */
   List<Partition> alterPartitions(final RawStore msdb, Warehouse wh, final String catName,
     final String dbname, final String name, final List<Partition> new_parts,
-    EnvironmentContext environmentContext, long txnId, String writeIdList, long writeId,
+    EnvironmentContext environmentContext,  String writeIdList, long writeId,
     IHMSHandler handler)
       throws InvalidOperationException, InvalidObjectException, AlreadyExistsException, MetaException;
 }
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/hive/blob/cdb32a7f/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java
index 719f001..e16bf7a 100644
--- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java
+++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java
@@ -90,7 +90,7 @@ public class HiveAlterHandler implements AlterHandler {
   @Override
   public void alterTable(RawStore msdb, Warehouse wh, String catName, String dbname,
       String name, Table newt, EnvironmentContext environmentContext,
-      IHMSHandler handler, long txnId, String writeIdList)
+      IHMSHandler handler, String writeIdList)
           throws InvalidOperationException, MetaException {
     catName = normalizeIdentifier(catName);
     name = name.toLowerCase();
@@ -142,7 +142,7 @@ public class HiveAlterHandler implements AlterHandler {
 
       // check if table with the new name already exists
       if (!newTblName.equals(name) || !newDbName.equals(dbname)) {
-        if (msdb.getTable(catName, newDbName, newTblName,  -1, null) != null) {
+        if (msdb.getTable(catName, newDbName, newTblName, null) != null) {
           throw new InvalidOperationException("new table " + newDbName
               + "." + newTblName + " already exists");
         }
@@ -152,7 +152,7 @@ public class HiveAlterHandler implements AlterHandler {
       msdb.openTransaction();
       // get old table
       // Note: we don't verify stats here; it's done below in alterTableUpdateTableColumnStats.
-      oldt = msdb.getTable(catName, dbname, name, -1, null);
+      oldt = msdb.getTable(catName, dbname, name, null);
       if (oldt == null) {
         throw new InvalidOperationException("table " +
             TableName.getQualified(catName, dbname, name) + " doesn't exist");
@@ -278,7 +278,7 @@ public class HiveAlterHandler implements AlterHandler {
             }
           }
           // Do not verify stats parameters on a partitioned table.
-          msdb.alterTable(catName, dbname, name, newt, -1, null);
+          msdb.alterTable(catName, dbname, name, newt, null);
           // alterPartition is only for changing the partition location in the table rename
           if (dataWasMoved) {
 
@@ -297,7 +297,7 @@ public class HiveAlterHandler implements AlterHandler {
                 partValues.add(part.getValues());
               }
               msdb.alterPartitions(catName, newDbName, newTblName, partValues,
-                  partBatch, newt.getWriteId(), txnId, writeIdList);
+                  partBatch, newt.getWriteId(), writeIdList);
             }
           }
 
@@ -306,11 +306,11 @@ public class HiveAlterHandler implements AlterHandler {
             newPartColStats.getStatsDesc().setDbName(newDbName);
             newPartColStats.getStatsDesc().setTableName(newTblName);
             msdb.updatePartitionColumnStatistics(newPartColStats, partColStats.getKey().getValues(),
-                txnId, writeIdList, newt.getWriteId());
+                writeIdList, newt.getWriteId());
           }
         } else {
           alterTableUpdateTableColumnStats(
-              msdb, oldt, newt, environmentContext, txnId, writeIdList);
+              msdb, oldt, newt, environmentContext, writeIdList);
         }
       } else {
         // operations other than table rename
@@ -334,24 +334,24 @@ public class HiveAlterHandler implements AlterHandler {
               assert(colStats == null);
               if (cascade) {
                 msdb.alterPartition(
-                    catName, dbname, name, part.getValues(), part, txnId, writeIdList);
+                    catName, dbname, name, part.getValues(), part, writeIdList);
               } else {
                 // update changed properties (stats)
                 oldPart.setParameters(part.getParameters());
                 msdb.alterPartition(
-                    catName, dbname, name, part.getValues(), oldPart, txnId, writeIdList);
+                    catName, dbname, name, part.getValues(), oldPart, writeIdList);
               }
             }
             // Don't validate table-level stats for a partitoned table.
-            msdb.alterTable(catName, dbname, name, newt, -1, null);
+            msdb.alterTable(catName, dbname, name, newt, null);
           } else {
             LOG.warn("Alter table not cascaded to partitions.");
             alterTableUpdateTableColumnStats(
-                msdb, oldt, newt, environmentContext, txnId, writeIdList);
+                msdb, oldt, newt, environmentContext, writeIdList);
           }
         } else {
           alterTableUpdateTableColumnStats(
-              msdb, oldt, newt, environmentContext, txnId, writeIdList);
+              msdb, oldt, newt, environmentContext, writeIdList);
         }
       }
 
@@ -431,13 +431,13 @@ public class HiveAlterHandler implements AlterHandler {
     EnvironmentContext environmentContext)
       throws InvalidOperationException, InvalidObjectException, AlreadyExistsException, MetaException {
     return alterPartition(msdb, wh, DEFAULT_CATALOG_NAME, dbname, name, part_vals, new_part,
-        environmentContext, null, -1, null);
+        environmentContext, null, null);
   }
 
   @Override
   public Partition alterPartition(RawStore msdb, Warehouse wh, String catName, String dbname,
       String name, List<String> part_vals, final Partition new_part,
-      EnvironmentContext environmentContext, IHMSHandler handler, long txnId, String validWriteIds)
+      EnvironmentContext environmentContext, IHMSHandler handler, String validWriteIds)
       throws InvalidOperationException, InvalidObjectException, AlreadyExistsException, MetaException {
     boolean success = false;
     Partition oldPart;
@@ -459,7 +459,7 @@ public class HiveAlterHandler implements AlterHandler {
       try {
         msdb.openTransaction();
 
-        Table tbl = msdb.getTable(catName, dbname, name,  -1, null);
+        Table tbl = msdb.getTable(catName, dbname, name, null);
         if (tbl == null) {
           throw new InvalidObjectException(
               "Unable to alter partition because table or database does not exist.");
@@ -481,7 +481,7 @@ public class HiveAlterHandler implements AlterHandler {
               oldPart.getSd().getCols(), tbl, new_part, null);
         }
         msdb.alterPartition(
-            catName, dbname, name, new_part.getValues(), new_part, txnId, validWriteIds);
+            catName, dbname, name, new_part.getValues(), new_part, validWriteIds);
         if (transactionalListeners != null && !transactionalListeners.isEmpty()) {
           MetaStoreListenerNotifier.notifyEvent(transactionalListeners,
                                                 EventMessage.EventType.ALTER_PARTITION,
@@ -516,7 +516,7 @@ public class HiveAlterHandler implements AlterHandler {
     Database db;
     try {
       msdb.openTransaction();
-      Table tbl = msdb.getTable(DEFAULT_CATALOG_NAME, dbname, name,  -1, null);
+      Table tbl = msdb.getTable(DEFAULT_CATALOG_NAME, dbname, name, null);
       if (tbl == null) {
         throw new InvalidObjectException(
             "Unable to alter partition because table or database does not exist.");
@@ -617,12 +617,12 @@ public class HiveAlterHandler implements AlterHandler {
       String newPartName = Warehouse.makePartName(tbl.getPartitionKeys(), new_part.getValues());
       ColumnStatistics cs = updateOrGetPartitionColumnStats(msdb, catName, dbname, name, oldPart.getValues(),
           oldPart.getSd().getCols(), tbl, new_part, null);
-      msdb.alterPartition(catName, dbname, name, part_vals, new_part, txnId, validWriteIds);
+      msdb.alterPartition(catName, dbname, name, part_vals, new_part, validWriteIds);
       if (cs != null) {
         cs.getStatsDesc().setPartName(newPartName);
         try {
           msdb.updatePartitionColumnStatistics(cs, new_part.getValues(),
-              txnId, validWriteIds, new_part.getWriteId());
+              validWriteIds, new_part.getWriteId());
         } catch (InvalidInputException iie) {
           throw new InvalidOperationException("Unable to update partition stats in table rename." + iie);
         } catch (NoSuchObjectException nsoe) {
@@ -668,7 +668,7 @@ public class HiveAlterHandler implements AlterHandler {
     EnvironmentContext environmentContext)
       throws InvalidOperationException, InvalidObjectException, AlreadyExistsException, MetaException {
     return alterPartitions(msdb, wh, DEFAULT_CATALOG_NAME, dbname, name, new_parts,
-        environmentContext, -1, null, -1, null);
+        environmentContext, null, -1, null);
   }
 
   @Override
@@ -676,7 +676,7 @@ public class HiveAlterHandler implements AlterHandler {
                                          final String dbname, final String name,
                                          final List<Partition> new_parts,
                                          EnvironmentContext environmentContext,
-                                         long txnId, String writeIdList, long writeId,
+                                         String writeIdList, long writeId,
                                          IHMSHandler handler)
       throws InvalidOperationException, InvalidObjectException, AlreadyExistsException, MetaException {
     List<Partition> oldParts = new ArrayList<>();
@@ -691,7 +691,7 @@ public class HiveAlterHandler implements AlterHandler {
       msdb.openTransaction();
 
       // Note: should we pass in write ID here? We only update stats on parts so probably not.
-      Table tbl = msdb.getTable(catName, dbname, name,  -1, null);
+      Table tbl = msdb.getTable(catName, dbname, name, null);
       if (tbl == null) {
         throw new InvalidObjectException(
             "Unable to alter partitions because table or database does not exist.");
@@ -726,7 +726,7 @@ public class HiveAlterHandler implements AlterHandler {
         }
       }
 
-      msdb.alterPartitions(catName, dbname, name, partValsList, new_parts, writeId, txnId, writeIdList);
+      msdb.alterPartitions(catName, dbname, name, partValsList, new_parts, writeId, writeIdList);
       Iterator<Partition> oldPartsIt = oldParts.iterator();
       for (Partition newPart : new_parts) {
         Partition oldPart;
@@ -795,7 +795,7 @@ public class HiveAlterHandler implements AlterHandler {
 
   @VisibleForTesting
   void alterTableUpdateTableColumnStats(RawStore msdb, Table oldTable, Table newTable,
-      EnvironmentContext ec, long txnId, String validWriteIds)
+      EnvironmentContext ec, String validWriteIds)
       throws MetaException, InvalidObjectException {
     String catName = normalizeIdentifier(oldTable.isSetCatName() ? oldTable.getCatName() :
         getDefaultCatalog(conf));
@@ -854,13 +854,13 @@ public class HiveAlterHandler implements AlterHandler {
       }
 
       // Change to new table and append stats for the new table
-      msdb.alterTable(catName, dbName, tableName, newTable, txnId, validWriteIds);
+      msdb.alterTable(catName, dbName, tableName, newTable, validWriteIds);
       if (updateColumnStats && !newStatsObjs.isEmpty()) {
         ColumnStatisticsDesc statsDesc = colStats.getStatsDesc();
         statsDesc.setDbName(newDbName);
         statsDesc.setTableName(newTableName);
         colStats.setStatsObj(newStatsObjs);
-        msdb.updateTableColumnStatistics(colStats, txnId, validWriteIds, newTable.getWriteId());
+        msdb.updateTableColumnStatistics(colStats, validWriteIds, newTable.getWriteId());
       }
     } catch (NoSuchObjectException nsoe) {
       LOG.debug("Could not find db entry." + nsoe);

http://git-wip-us.apache.org/repos/asf/hive/blob/cdb32a7f/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java
index 285f7fb..eb4eb1b 100644
--- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java
+++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java
@@ -2435,7 +2435,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
 
     private boolean is_table_exists(RawStore ms, String catName, String dbname, String name)
         throws MetaException {
-      return (ms.getTable(catName, dbname, name, -1, null) != null);
+      return (ms.getTable(catName, dbname, name, null) != null);
     }
 
     private boolean drop_table_core(final RawStore ms, final String catName, final String dbname,
@@ -2694,7 +2694,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
     }
 
     private void alterPartitionForTruncate(RawStore ms, String catName, String dbName, String tableName,
-        Table table, Partition partition, long txnId, String validWriteIds, long writeId) throws Exception {
+        Table table, Partition partition, String validWriteIds, long writeId) throws Exception {
       EnvironmentContext environmentContext = new EnvironmentContext();
       updateStatsForTruncate(partition.getParameters(), environmentContext);
 
@@ -2714,17 +2714,17 @@ public class HiveMetaStore extends ThriftHiveMetastore {
         partition.setWriteId(writeId);
       }
       alterHandler.alterPartition(ms, wh, catName, dbName, tableName, null, partition,
-          environmentContext, this, txnId, validWriteIds);
+          environmentContext, this, validWriteIds);
     }
 
     private void alterTableStatsForTruncate(RawStore ms, String catName, String dbName,
         String tableName, Table table, List<String> partNames,
-        long txnId, String validWriteIds, long writeId) throws Exception {
+        String validWriteIds, long writeId) throws Exception {
       if (partNames == null) {
         if (0 != table.getPartitionKeysSize()) {
           for (Partition partition : ms.getPartitions(catName, dbName, tableName, Integer.MAX_VALUE)) {
             alterPartitionForTruncate(ms, catName, dbName, tableName, table, partition,
-                txnId, validWriteIds, writeId);
+                validWriteIds, writeId);
           }
         } else {
           EnvironmentContext environmentContext = new EnvironmentContext();
@@ -2747,12 +2747,12 @@ public class HiveMetaStore extends ThriftHiveMetastore {
             table.setWriteId(writeId);
           }
           alterHandler.alterTable(ms, wh, catName, dbName, tableName, table,
-              environmentContext, this, txnId, validWriteIds);
+              environmentContext, this, validWriteIds);
         }
       } else {
         for (Partition partition : ms.getPartitionsByNames(catName, dbName, tableName, partNames)) {
           alterPartitionForTruncate(ms, catName, dbName, tableName, table, partition,
-              txnId, validWriteIds, writeId);
+              validWriteIds, writeId);
         }
       }
       return;
@@ -2791,19 +2791,19 @@ public class HiveMetaStore extends ThriftHiveMetastore {
     public void truncate_table(final String dbName, final String tableName, List<String> partNames)
       throws NoSuchObjectException, MetaException {
       // Deprecated path, won't work for txn tables.
-      truncateTableInternal(dbName, tableName, partNames, -1, null, -1);
+      truncateTableInternal(dbName, tableName, partNames, null, -1);
     }
 
     @Override
     public TruncateTableResponse truncate_table_req(TruncateTableRequest req)
         throws MetaException, TException {
       truncateTableInternal(req.getDbName(), req.getTableName(), req.getPartNames(),
-          req.getTxnId(), req.getValidWriteIdList(), req.getWriteId());
+          req.getValidWriteIdList(), req.getWriteId());
       return new TruncateTableResponse();
     }
 
     private void truncateTableInternal(String dbName, String tableName, List<String> partNames,
-        long txnId, String validWriteIds, long writeId) throws MetaException, NoSuchObjectException {
+        String validWriteIds, long writeId) throws MetaException, NoSuchObjectException {
       try {
         String[] parsedDbName = parseDbName(dbName, conf);
         Table tbl = get_table_core(parsedDbName[CAT_NAME], parsedDbName[DB_NAME], tableName);
@@ -2835,7 +2835,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
 
         // Alter the table/partition stats and also notify truncate table event
         alterTableStatsForTruncate(getMS(), parsedDbName[CAT_NAME], parsedDbName[DB_NAME],
-            tableName, tbl, partNames, txnId, validWriteIds, writeId);
+            tableName, tbl, partNames, validWriteIds, writeId);
       } catch (IOException e) {
         throw new MetaException(e.getMessage());
       } catch (MetaException | NoSuchObjectException e) {
@@ -2866,7 +2866,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
         NoSuchObjectException {
       String[] parsedDbName = parseDbName(dbname, conf);
       return getTableInternal(
-          parsedDbName[CAT_NAME], parsedDbName[DB_NAME], name, null, -1, null);
+          parsedDbName[CAT_NAME], parsedDbName[DB_NAME], name, null, null);
     }
 
     @Override
@@ -2874,11 +2874,11 @@ public class HiveMetaStore extends ThriftHiveMetastore {
         NoSuchObjectException {
       String catName = req.isSetCatName() ? req.getCatName() : getDefaultCatalog(conf);
       return new GetTableResult(getTableInternal(catName, req.getDbName(), req.getTblName(),
-          req.getCapabilities(), req.getTxnId(), req.getValidWriteIdList()));
+          req.getCapabilities(), req.getValidWriteIdList()));
     }
 
     private Table getTableInternal(String catName, String dbname, String name,
-        ClientCapabilities capabilities, long txnId, String writeIdList)
+        ClientCapabilities capabilities, String writeIdList)
         throws MetaException, NoSuchObjectException {
       if (isInTest) {
         assertClientHasCapability(capabilities, ClientCapability.TEST_CAPABILITY,
@@ -2889,7 +2889,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
       startTableFunction("get_table", catName, dbname, name);
       Exception ex = null;
       try {
-        t = get_table_core(catName, dbname, name, txnId, writeIdList);
+        t = get_table_core(catName, dbname, name, writeIdList);
         if (MetaStoreUtils.isInsertOnlyTableParam(t.getParameters())) {
           assertClientHasCapability(capabilities, ClientCapability.INSERT_ONLY_TABLES,
               "insert-only tables", "get_table_req");
@@ -2929,7 +2929,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
         final String dbname,
         final String name)
         throws MetaException, NoSuchObjectException {
-      return get_table_core(catName, dbname, name, -1, null);
+      return get_table_core(catName, dbname, name, null);
     }
 
     @Override
@@ -2937,12 +2937,11 @@ public class HiveMetaStore extends ThriftHiveMetastore {
         final String catName,
         final String dbname,
         final String name,
-        final long txnId,
         final String writeIdList)
         throws MetaException, NoSuchObjectException {
       Table t = null;
       try {
-        t = getMS().getTable(catName, dbname, name, txnId, writeIdList);
+        t = getMS().getTable(catName, dbname, name, writeIdList);
         if (t == null) {
           throw new NoSuchObjectException(TableName.getQualified(catName, dbname, name) +
             " table not found");
@@ -3124,7 +3123,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
 
         MetaStoreUtils.validatePartitionNameCharacters(part_vals, partitionValidationPattern);
 
-        tbl = ms.getTable(part.getCatName(), part.getDbName(), part.getTableName(), -1, null);
+        tbl = ms.getTable(part.getCatName(), part.getDbName(), part.getTableName(), null);
         if (tbl == null) {
           throw new InvalidObjectException(
               "Unable to add partition because table or database do not exist");
@@ -3318,7 +3317,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
 
       try {
         ms.openTransaction();
-        tbl = ms.getTable(catName, dbName, tblName, -1, null);
+        tbl = ms.getTable(catName, dbName, tblName, null);
         if (tbl == null) {
           throw new InvalidObjectException("Unable to add partitions because "
               + TableName.getQualified(catName, dbName, tblName) +
@@ -3674,7 +3673,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
       Database db = null;
       try {
         ms.openTransaction();
-        tbl = ms.getTable(catName, dbName, tblName, -1, null);
+        tbl = ms.getTable(catName, dbName, tblName, null);
         if (tbl == null) {
           throw new InvalidObjectException("Unable to add partitions because "
               + "database or table " + dbName + "." + tblName + " does not exist");
@@ -3844,7 +3843,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
       }
       try {
         ms.openTransaction();
-        tbl = ms.getTable(part.getCatName(), part.getDbName(), part.getTableName(), -1, null);
+        tbl = ms.getTable(part.getCatName(), part.getDbName(), part.getTableName(), null);
         if (tbl == null) {
           throw new InvalidObjectException(
               "Unable to add partition because table or database do not exist");
@@ -3966,7 +3965,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
 
       Table destinationTable =
           ms.getTable(
-              parsedDestDbName[CAT_NAME], parsedDestDbName[DB_NAME], destTableName, -1, null);
+              parsedDestDbName[CAT_NAME], parsedDestDbName[DB_NAME], destTableName, null);
       if (destinationTable == null) {
         throw new MetaException( "The destination table " +
             TableName.getQualified(parsedDestDbName[CAT_NAME],
@@ -3974,7 +3973,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
       }
       Table sourceTable =
           ms.getTable(
-              parsedSourceDbName[CAT_NAME], parsedSourceDbName[DB_NAME], sourceTableName, -1, null);
+              parsedSourceDbName[CAT_NAME], parsedSourceDbName[DB_NAME], sourceTableName, null);
       if (sourceTable == null) {
         throw new MetaException("The source table " +
             TableName.getQualified(parsedSourceDbName[CAT_NAME],
@@ -4151,7 +4150,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
       try {
         ms.openTransaction();
         part = ms.getPartition(catName, db_name, tbl_name, part_vals);
-        tbl = get_table_core(catName, db_name, tbl_name, -1, null);
+        tbl = get_table_core(catName, db_name, tbl_name, null);
         tableDataShouldBeDeleted = checkTableDataShouldBeDeleted(tbl, deleteData);
         firePreEvent(new PreDropPartitionEvent(tbl, part, deleteData, this));
         mustPurge = isMustPurge(envContext, tbl);
@@ -4826,7 +4825,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
       String[] parsedDbName = parseDbName(dbName, conf);
       // TODO: this method name is confusing, it actually does full alter (sortof)
       rename_partition(parsedDbName[CAT_NAME], parsedDbName[DB_NAME], tableName, null, newPartition,
-          envContext, -1, null);
+          envContext, null);
     }
 
     @Deprecated
@@ -4837,18 +4836,18 @@ public class HiveMetaStore extends ThriftHiveMetastore {
       // Call rename_partition without an environment context.
       String[] parsedDbName = parseDbName(db_name, conf);
       rename_partition(parsedDbName[CAT_NAME], parsedDbName[DB_NAME], tbl_name, part_vals, new_part,
-          null, -1, null);
+          null, null);
     }
 
     public RenamePartitionResponse rename_partition_req(
         RenamePartitionRequest req) throws InvalidOperationException ,MetaException ,TException {
       rename_partition(req.getCatName(), req.getDbName(), req.getTableName(), req.getPartVals(),
-          req.getNewPart(), null, req.getTxnId(), req.getValidWriteIdList());
+          req.getNewPart(), null, req.getValidWriteIdList());
       return new RenamePartitionResponse();
     };
 
     private void rename_partition(String catName, String db_name, String tbl_name,
-        List<String> part_vals, Partition new_part, EnvironmentContext envContext, long txnId,
+        List<String> part_vals, Partition new_part, EnvironmentContext envContext,
         String validWriteIds) throws TException {
       startTableFunction("alter_partition", catName, db_name, tbl_name);
 
@@ -4883,13 +4882,13 @@ public class HiveMetaStore extends ThriftHiveMetastore {
         }
 
         oldPart = alterHandler.alterPartition(getMS(), wh, catName, db_name, tbl_name,
-            part_vals, new_part, envContext, this, txnId, validWriteIds);
+            part_vals, new_part, envContext, this, validWriteIds);
 
         // Only fetch the table if we actually have a listener
         Table table = null;
         if (!listeners.isEmpty()) {
           if (table == null) {
-            table = getMS().getTable(catName, db_name, tbl_name, -1, null);
+            table = getMS().getTable(catName, db_name, tbl_name, null);
           }
 
           MetaStoreListenerNotifier.notifyEvent(listeners,
@@ -4920,14 +4919,13 @@ public class HiveMetaStore extends ThriftHiveMetastore {
         throws TException {
       String[] o = parseDbName(db_name, conf);
       alter_partitions_with_environment_context(o[0], o[1],
-          tbl_name, new_parts, null, -1, null, -1);
+          tbl_name, new_parts, null, null, -1);
     }
 
     @Override
     public AlterPartitionsResponse alter_partitions_req(AlterPartitionsRequest req) throws TException {
       alter_partitions_with_environment_context(req.getCatName(),
           req.getDbName(), req.getTableName(), req.getPartitions(), req.getEnvironmentContext(),
-          req.isSetTxnId() ? req.getTxnId() : -1,
           req.isSetValidWriteIdList() ? req.getValidWriteIdList() : null,
           req.isSetWriteId() ? req.getWriteId() : -1);
       return new AlterPartitionsResponse();
@@ -4941,12 +4939,12 @@ public class HiveMetaStore extends ThriftHiveMetastore {
             throws TException {
       String[] o = parseDbName(db_name, conf);
       alter_partitions_with_environment_context(o[0], o[1], tbl_name, new_parts, environmentContext,
-          -1, null, -1);
+          null, -1);
     }
 
     private void alter_partitions_with_environment_context(String catName, String db_name, final String tbl_name,
         final List<Partition> new_parts, EnvironmentContext environmentContext,
-        long txnId, String writeIdList, long writeId)
+        String writeIdList, long writeId)
         throws TException {
       if (environmentContext == null) {
         environmentContext = new EnvironmentContext();
@@ -4975,7 +4973,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
           firePreEvent(new PreAlterPartitionEvent(db_name, tbl_name, null, tmpPart, this));
         }
         oldParts = alterHandler.alterPartitions(getMS(), wh,
-            catName, db_name, tbl_name, new_parts, environmentContext, txnId, writeIdList, writeId, this);
+            catName, db_name, tbl_name, new_parts, environmentContext, writeIdList, writeId, this);
         Iterator<Partition> olditr = oldParts.iterator();
         // Only fetch the table if we have a listener that needs it.
         Table table = null;
@@ -4989,7 +4987,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
           }
 
           if (table == null) {
-            table = getMS().getTable(catName, db_name, tbl_name,  -1, null);
+            table = getMS().getTable(catName, db_name, tbl_name,  null);
           }
 
           if (!listeners.isEmpty()) {
@@ -5028,7 +5026,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
       // Do not set an environment context.
       String[] parsedDbName = parseDbName(dbname, conf);
       alter_table_core(parsedDbName[CAT_NAME], parsedDbName[DB_NAME], name, newTable,
-          null, -1, null);
+          null, null);
     }
 
     @Override
@@ -5042,14 +5040,14 @@ public class HiveMetaStore extends ThriftHiveMetastore {
       }
       String[] parsedDbName = parseDbName(dbname, conf);
       alter_table_core(parsedDbName[CAT_NAME], parsedDbName[DB_NAME], name, newTable,
-          envContext, -1, null);
+          envContext, null);
     }
 
     @Override
     public AlterTableResponse alter_table_req(AlterTableRequest req)
         throws InvalidOperationException, MetaException, TException {
       alter_table_core(req.getCatName(), req.getDbName(), req.getTableName(),
-          req.getTable(), req.getEnvironmentContext(), req.getTxnId(), req.getValidWriteIdList());
+          req.getTable(), req.getEnvironmentContext(), req.getValidWriteIdList());
       return new AlterTableResponse();
     }
 
@@ -5060,11 +5058,11 @@ public class HiveMetaStore extends ThriftHiveMetastore {
         throws InvalidOperationException, MetaException {
       String[] parsedDbName = parseDbName(dbname, conf);
       alter_table_core(parsedDbName[CAT_NAME], parsedDbName[DB_NAME],
-          name, newTable, envContext, -1, null);
+          name, newTable, envContext, null);
     }
 
     private void alter_table_core(String catName, String dbname, String name, Table newTable,
-        EnvironmentContext envContext, long txnId, String validWriteIdList)
+        EnvironmentContext envContext, String validWriteIdList)
         throws InvalidOperationException, MetaException {
       startFunction("alter_table", ": " + TableName.getQualified(catName, dbname, name)
           + " newtbl=" + newTable.getTableName());
@@ -5101,7 +5099,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
         Table oldt = get_table_core(catName, dbname, name);
         firePreEvent(new PreAlterTableEvent(oldt, newTable, this));
         alterHandler.alterTable(getMS(), wh, catName, dbname, name, newTable,
-                envContext, this, txnId, validWriteIdList);
+                envContext, this, validWriteIdList);
         success = true;
       } catch (NoSuchObjectException e) {
         // thrown when the table to be altered does not exist
@@ -5426,7 +5424,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
     private List<String> getPartValsFromName(RawStore ms, String catName, String dbName,
                                              String tblName, String partName)
         throws MetaException, InvalidObjectException {
-      Table t = ms.getTable(catName, dbName, tblName,  -1, null);
+      Table t = ms.getTable(catName, dbName, tblName,  null);
       if (t == null) {
         throw new InvalidObjectException(dbName + "." + tblName
             + " table not found");
@@ -5684,7 +5682,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
       try {
         statsObj = getMS().getTableColumnStatistics(
             parsedDbName[CAT_NAME], parsedDbName[DB_NAME], tableName, Lists.newArrayList(colName),
-            -1, null);
+            null);
         if (statsObj != null) {
           assert statsObj.getStatsObjSize() <= 1;
         }
@@ -5710,7 +5708,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
       try {
         ColumnStatistics cs = getMS().getTableColumnStatistics(
             catName, dbName, tblName, lowerCaseColNames,
-            request.getTxnId(), request.getValidWriteIdList());
+            request.getValidWriteIdList());
         // Note: stats compliance is not propagated to the client; instead, we just return nothing
         //       if stats are not compliant for now. This won't work for stats merging, but that
         //       is currently only done on metastore size (see set_aggr...).
@@ -5779,7 +5777,6 @@ public class HiveMetaStore extends ThriftHiveMetastore {
       try {
         List<ColumnStatistics> stats = getMS().getPartitionColumnStatistics(
             catName, dbName, tblName, lowerCasePartNames, lowerCaseColNames,
-            request.isSetTxnId() ? request.getTxnId() : -1,
             request.isSetValidWriteIdList() ? request.getValidWriteIdList() : null);
         Map<String, List<ColumnStatisticsObj>> map = new HashMap<>();
         if (stats != null) {
@@ -5803,7 +5800,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
     @Override
     public boolean update_table_column_statistics(ColumnStatistics colStats) throws TException {
       // Deprecated API, won't work for transactional tables
-      return updateTableColumnStatsInternal(colStats, -1, null, -1);
+      return updateTableColumnStatsInternal(colStats, null, -1);
     }
 
     @Override
@@ -5819,12 +5816,12 @@ public class HiveMetaStore extends ThriftHiveMetastore {
       }
       ColumnStatistics colStats = req.getColStatsIterator().next();
       boolean ret = updateTableColumnStatsInternal(colStats,
-          req.getTxnId(), req.getValidWriteIdList(), req.getWriteId());
+          req.getValidWriteIdList(), req.getWriteId());
       return new SetPartitionsStatsResponse(ret);
     }
 
     private boolean updateTableColumnStatsInternal(ColumnStatistics colStats,
-        long txnId, String validWriteIds, long writeId)
+        String validWriteIds, long writeId)
         throws NoSuchObjectException, MetaException, InvalidObjectException, InvalidInputException {
       normalizeColStatsInput(colStats);
 
@@ -5834,7 +5831,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
 
       boolean ret = false;
       try {
-        ret = getMS().updateTableColumnStatistics(colStats, txnId, validWriteIds, writeId);
+        ret = getMS().updateTableColumnStatistics(colStats, validWriteIds, writeId);
       } finally {
         endFunction("write_column_statistics", ret != false, null,
             colStats.getStatsDesc().getTableName());
@@ -5861,7 +5858,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
     }
 
     private boolean updatePartitonColStatsInternal(Table tbl, ColumnStatistics colStats,
-        long txnId, String validWriteIds, long writeId)
+        String validWriteIds, long writeId)
         throws MetaException, InvalidObjectException, NoSuchObjectException, InvalidInputException {
       normalizeColStatsInput(colStats);
 
@@ -5878,7 +5875,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
         }
         List<String> partVals = getPartValsFromName(tbl, csd.getPartName());
         return getMS().updatePartitionColumnStatistics(
-            colStats, partVals, txnId, validWriteIds, writeId);
+            colStats, partVals, validWriteIds, writeId);
       } finally {
         endFunction("write_partition_column_statistics", ret != false, null, tableName);
       }
@@ -5887,7 +5884,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
     @Override
     public boolean update_partition_column_statistics(ColumnStatistics colStats) throws TException {
       // Deprecated API.
-      return updatePartitonColStatsInternal(null, colStats, -1, null, -1);
+      return updatePartitonColStatsInternal(null, colStats, null, -1);
     }
 
 
@@ -5904,7 +5901,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
       }
       ColumnStatistics colStats = req.getColStatsIterator().next();
       boolean ret = updatePartitonColStatsInternal(null, colStats,
-          req.getTxnId(), req.getValidWriteIdList(), req.getWriteId());
+          req.getValidWriteIdList(), req.getWriteId());
       return new SetPartitionsStatsResponse(ret);
     }
 
@@ -7560,8 +7557,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
 
       try {
         aggrStats = getMS().get_aggr_stats_for(catName, dbName, tblName,
-            lowerCasePartNames, lowerCaseColNames, request.getTxnId(),
-            request.getValidWriteIdList());
+            lowerCasePartNames, lowerCaseColNames, request.getValidWriteIdList());
         return aggrStats;
       } finally {
           endFunction("get_aggr_stats_for", aggrStats == null, null, request.getTblName());
@@ -7597,12 +7593,11 @@ public class HiveMetaStore extends ThriftHiveMetastore {
           return updateTableColumnStatsWithMerge(catName, dbName, tableName, colNames, request);
         } else {
           // This is the overwrite case, we do not care about the accuracy.
-          return updateTableColumnStatsInternal(firstColStats, request.getTxnId(),
+          return updateTableColumnStatsInternal(firstColStats,
               request.getValidWriteIdList(), request.getWriteId());
         }
       } else {
         // partition level column stats merging
-        List<Partition> partitions = new ArrayList<>();
         // note that we may have two or more duplicate partition names.
         // see autoColumnStats_2.q under TestMiniLlapLocalCliDriver
         Map<String, ColumnStatistics> newStatsMap = new HashMap<>();
@@ -7621,7 +7616,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
           Table t = getTable(catName, dbName, tableName);
           for (Entry<String, ColumnStatistics> entry : newStatsMap.entrySet()) {
             // We don't short-circuit on errors here anymore. That can leave acid stats invalid.
-            ret = updatePartitonColStatsInternal(t, entry.getValue(), request.getTxnId(),
+            ret = updatePartitonColStatsInternal(t, entry.getValue(),
                 request.getValidWriteIdList(), request.getWriteId()) && ret;
           }
         }
@@ -7640,7 +7635,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
         List<String> partitionNames = new ArrayList<>();
         partitionNames.addAll(newStatsMap.keySet());
         List<ColumnStatistics> csOlds = ms.getPartitionColumnStatistics(catName, dbName, tableName,
-            partitionNames, colNames, request.getTxnId(), request.getValidWriteIdList());
+            partitionNames, colNames, request.getValidWriteIdList());
         if (newStatsMap.values().size() != csOlds.size()) {
           // some of the partitions miss stats.
           LOG.debug("Some of the partitions miss stats.");
@@ -7678,7 +7673,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
 
           if (!csNew.getStatsObj().isEmpty()) {
             // We don't short-circuit on errors here anymore. That can leave acid stats invalid.
-            result = updatePartitonColStatsInternal(t, csNew, request.getTxnId(),
+            result = updatePartitonColStatsInternal(t, csNew,
                 request.getValidWriteIdList(), request.getWriteId()) && result;
           } else if (isInvalidTxnStats) {
             // For now because the stats state is such as it is, we will invalidate everything.
@@ -7688,7 +7683,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
             StatsSetupConst.clearColumnStatsState(part.getParameters());
             StatsSetupConst.setBasicStatsState(part.getParameters(), StatsSetupConst.FALSE);
             ms.alterPartition(catName, dbName, tableName, part.getValues(), part,
-                request.getTxnId(), request.getValidWriteIdList());
+                request.getValidWriteIdList());
             result = false;
           } else {
             // TODO: why doesn't the original call for non acid tables invalidate the stats?
@@ -7716,7 +7711,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
       boolean isCommitted = false, result = false;
       try {
         ColumnStatistics csOld = ms.getTableColumnStatistics(catName, dbName, tableName, colNames,
-            request.getTxnId(), request.getValidWriteIdList());
+            request.getValidWriteIdList());
         // we first use the valid stats list to prune the stats
         boolean isInvalidTxnStats = csOld != null
             && csOld.isSetIsStatsCompliant() && !csOld.isIsStatsCompliant();
@@ -7734,7 +7729,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
         }
 
         if (!firstColStats.getStatsObj().isEmpty()) {
-          result = updateTableColumnStatsInternal(firstColStats, request.getTxnId(),
+          result = updateTableColumnStatsInternal(firstColStats,
               request.getValidWriteIdList(), request.getWriteId());
         } else if (isInvalidTxnStats) {
           // For now because the stats state is such as it is, we will invalidate everything.
@@ -7744,7 +7739,7 @@ public class HiveMetaStore extends ThriftHiveMetastore {
           t.setWriteId(request.getWriteId());
           StatsSetupConst.clearColumnStatsState(t.getParameters());
           StatsSetupConst.setBasicStatsState(t.getParameters(), StatsSetupConst.FALSE);
-          ms.alterTable(catName, dbName, tableName, t, request.getTxnId(), request.getValidWriteIdList());
+          ms.alterTable(catName, dbName, tableName, t, request.getValidWriteIdList());
         } else {
           // TODO: why doesn't the original call for non acid tables invalidate the stats?
           LOG.debug("All the column stats are not accurate to merge.");
@@ -7763,13 +7758,13 @@ public class HiveMetaStore extends ThriftHiveMetastore {
 
     private Table getTable(String catName, String dbName, String tableName)
         throws MetaException, InvalidObjectException {
-      return getTable(catName, dbName, tableName, -1, null);
+      return getTable(catName, dbName, tableName, null);
     }
 
     private Table getTable(String catName, String dbName, String tableName,
-                           long txnId, String writeIdList)
+                           String writeIdList)
         throws MetaException, InvalidObjectException {
-      Table t = getMS().getTable(catName, dbName, tableName, txnId, writeIdList);
+      Table t = getMS().getTable(catName, dbName, tableName, writeIdList);
       if (t == null) {
         throw new InvalidObjectException(TableName.getQualified(catName, dbName, tableName)
             + " table not found");