You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by se...@apache.org on 2018/07/03 17:48:07 UTC

[39/46] hive git commit: HIVE-19416 : merge master into branch (Sergey Shelukhin) 0703

http://git-wip-us.apache.org/repos/asf/hive/blob/4db8b1c6/standalone-metastore/src/gen/thrift/gen-py/hive_metastore/ttypes.py
----------------------------------------------------------------------
diff --cc standalone-metastore/src/gen/thrift/gen-py/hive_metastore/ttypes.py
index 9b02951,031e72b..6e0e598
--- a/standalone-metastore/src/gen/thrift/gen-py/hive_metastore/ttypes.py
+++ b/standalone-metastore/src/gen/thrift/gen-py/hive_metastore/ttypes.py
@@@ -22074,213 -22043,6 +22463,213 @@@ class GetRuntimeStatsRequest
      value = (value * 31) ^ hash(self.maxWeight)
      value = (value * 31) ^ hash(self.maxCreateTime)
      return value
 +
 +  def __repr__(self):
 +    L = ['%s=%r' % (key, value)
 +      for key, value in self.__dict__.iteritems()]
 +    return '%s(%s)' % (self.__class__.__name__, ', '.join(L))
 +
 +  def __eq__(self, other):
 +    return isinstance(other, self.__class__) and self.__dict__ == other.__dict__
 +
 +  def __ne__(self, other):
 +    return not (self == other)
 +
 +class AlterPartitionsRequest:
 +  """
 +  Attributes:
 +   - dbName
 +   - tableName
 +   - partitions
 +   - environmentContext
 +   - txnId
 +   - writeId
 +   - validWriteIdList
 +  """
 +
 +  thrift_spec = (
 +    None, # 0
 +    (1, TType.STRING, 'dbName', None, None, ), # 1
 +    (2, TType.STRING, 'tableName', None, None, ), # 2
 +    (3, TType.LIST, 'partitions', (TType.STRUCT,(Partition, Partition.thrift_spec)), None, ), # 3
 +    (4, TType.STRUCT, 'environmentContext', (EnvironmentContext, EnvironmentContext.thrift_spec), None, ), # 4
 +    (5, TType.I64, 'txnId', None, -1, ), # 5
 +    (6, TType.I64, 'writeId', None, -1, ), # 6
 +    (7, TType.STRING, 'validWriteIdList', None, None, ), # 7
 +  )
 +
 +  def __init__(self, dbName=None, tableName=None, partitions=None, environmentContext=None, txnId=thrift_spec[5][4], writeId=thrift_spec[6][4], validWriteIdList=None,):
 +    self.dbName = dbName
 +    self.tableName = tableName
 +    self.partitions = partitions
 +    self.environmentContext = environmentContext
 +    self.txnId = txnId
 +    self.writeId = writeId
 +    self.validWriteIdList = validWriteIdList
 +
 +  def read(self, iprot):
 +    if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None:
 +      fastbinary.decode_binary(self, iprot.trans, (self.__class__, self.thrift_spec))
 +      return
 +    iprot.readStructBegin()
 +    while True:
 +      (fname, ftype, fid) = iprot.readFieldBegin()
 +      if ftype == TType.STOP:
 +        break
 +      if fid == 1:
 +        if ftype == TType.STRING:
 +          self.dbName = iprot.readString()
 +        else:
 +          iprot.skip(ftype)
 +      elif fid == 2:
 +        if ftype == TType.STRING:
 +          self.tableName = iprot.readString()
 +        else:
 +          iprot.skip(ftype)
 +      elif fid == 3:
 +        if ftype == TType.LIST:
 +          self.partitions = []
-           (_etype819, _size816) = iprot.readListBegin()
-           for _i820 in xrange(_size816):
-             _elem821 = Partition()
-             _elem821.read(iprot)
-             self.partitions.append(_elem821)
++          (_etype840, _size837) = iprot.readListBegin()
++          for _i841 in xrange(_size837):
++            _elem842 = Partition()
++            _elem842.read(iprot)
++            self.partitions.append(_elem842)
 +          iprot.readListEnd()
 +        else:
 +          iprot.skip(ftype)
 +      elif fid == 4:
 +        if ftype == TType.STRUCT:
 +          self.environmentContext = EnvironmentContext()
 +          self.environmentContext.read(iprot)
 +        else:
 +          iprot.skip(ftype)
 +      elif fid == 5:
 +        if ftype == TType.I64:
 +          self.txnId = iprot.readI64()
 +        else:
 +          iprot.skip(ftype)
 +      elif fid == 6:
 +        if ftype == TType.I64:
 +          self.writeId = iprot.readI64()
 +        else:
 +          iprot.skip(ftype)
 +      elif fid == 7:
 +        if ftype == TType.STRING:
 +          self.validWriteIdList = iprot.readString()
 +        else:
 +          iprot.skip(ftype)
 +      else:
 +        iprot.skip(ftype)
 +      iprot.readFieldEnd()
 +    iprot.readStructEnd()
 +
 +  def write(self, oprot):
 +    if oprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and self.thrift_spec is not None and fastbinary is not None:
 +      oprot.trans.write(fastbinary.encode_binary(self, (self.__class__, self.thrift_spec)))
 +      return
 +    oprot.writeStructBegin('AlterPartitionsRequest')
 +    if self.dbName is not None:
 +      oprot.writeFieldBegin('dbName', TType.STRING, 1)
 +      oprot.writeString(self.dbName)
 +      oprot.writeFieldEnd()
 +    if self.tableName is not None:
 +      oprot.writeFieldBegin('tableName', TType.STRING, 2)
 +      oprot.writeString(self.tableName)
 +      oprot.writeFieldEnd()
 +    if self.partitions is not None:
 +      oprot.writeFieldBegin('partitions', TType.LIST, 3)
 +      oprot.writeListBegin(TType.STRUCT, len(self.partitions))
-       for iter822 in self.partitions:
-         iter822.write(oprot)
++      for iter843 in self.partitions:
++        iter843.write(oprot)
 +      oprot.writeListEnd()
 +      oprot.writeFieldEnd()
 +    if self.environmentContext is not None:
 +      oprot.writeFieldBegin('environmentContext', TType.STRUCT, 4)
 +      self.environmentContext.write(oprot)
 +      oprot.writeFieldEnd()
 +    if self.txnId is not None:
 +      oprot.writeFieldBegin('txnId', TType.I64, 5)
 +      oprot.writeI64(self.txnId)
 +      oprot.writeFieldEnd()
 +    if self.writeId is not None:
 +      oprot.writeFieldBegin('writeId', TType.I64, 6)
 +      oprot.writeI64(self.writeId)
 +      oprot.writeFieldEnd()
 +    if self.validWriteIdList is not None:
 +      oprot.writeFieldBegin('validWriteIdList', TType.STRING, 7)
 +      oprot.writeString(self.validWriteIdList)
 +      oprot.writeFieldEnd()
 +    oprot.writeFieldStop()
 +    oprot.writeStructEnd()
 +
 +  def validate(self):
 +    if self.dbName is None:
 +      raise TProtocol.TProtocolException(message='Required field dbName is unset!')
 +    if self.tableName is None:
 +      raise TProtocol.TProtocolException(message='Required field tableName is unset!')
 +    if self.partitions is None:
 +      raise TProtocol.TProtocolException(message='Required field partitions is unset!')
 +    if self.environmentContext is None:
 +      raise TProtocol.TProtocolException(message='Required field environmentContext is unset!')
 +    return
 +
 +
 +  def __hash__(self):
 +    value = 17
 +    value = (value * 31) ^ hash(self.dbName)
 +    value = (value * 31) ^ hash(self.tableName)
 +    value = (value * 31) ^ hash(self.partitions)
 +    value = (value * 31) ^ hash(self.environmentContext)
 +    value = (value * 31) ^ hash(self.txnId)
 +    value = (value * 31) ^ hash(self.writeId)
 +    value = (value * 31) ^ hash(self.validWriteIdList)
 +    return value
 +
 +  def __repr__(self):
 +    L = ['%s=%r' % (key, value)
 +      for key, value in self.__dict__.iteritems()]
 +    return '%s(%s)' % (self.__class__.__name__, ', '.join(L))
 +
 +  def __eq__(self, other):
 +    return isinstance(other, self.__class__) and self.__dict__ == other.__dict__
 +
 +  def __ne__(self, other):
 +    return not (self == other)
 +
 +class AlterPartitionsResponse:
 +
 +  thrift_spec = (
 +  )
 +
 +  def read(self, iprot):
 +    if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None:
 +      fastbinary.decode_binary(self, iprot.trans, (self.__class__, self.thrift_spec))
 +      return
 +    iprot.readStructBegin()
 +    while True:
 +      (fname, ftype, fid) = iprot.readFieldBegin()
 +      if ftype == TType.STOP:
 +        break
 +      else:
 +        iprot.skip(ftype)
 +      iprot.readFieldEnd()
 +    iprot.readStructEnd()
 +
 +  def write(self, oprot):
 +    if oprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and self.thrift_spec is not None and fastbinary is not None:
 +      oprot.trans.write(fastbinary.encode_binary(self, (self.__class__, self.thrift_spec)))
 +      return
 +    oprot.writeStructBegin('AlterPartitionsResponse')
 +    oprot.writeFieldStop()
 +    oprot.writeStructEnd()
 +
 +  def validate(self):
 +    return
 +
 +
 +  def __hash__(self):
 +    value = 17
 +    return value
  
    def __repr__(self):
      L = ['%s=%r' % (key, value)

http://git-wip-us.apache.org/repos/asf/hive/blob/4db8b1c6/standalone-metastore/src/gen/thrift/gen-rb/hive_metastore_types.rb
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/hive/blob/4db8b1c6/standalone-metastore/src/gen/thrift/gen-rb/thrift_hive_metastore.rb
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/hive/blob/4db8b1c6/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/hive/blob/4db8b1c6/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/hive/blob/4db8b1c6/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/IMetaStoreClient.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/hive/blob/4db8b1c6/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/ObjectStore.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/hive/blob/4db8b1c6/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/RawStore.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/hive/blob/4db8b1c6/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/cache/CachedStore.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/hive/blob/4db8b1c6/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnDbUtil.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/hive/blob/4db8b1c6/standalone-metastore/src/main/resources/package.jdo
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/hive/blob/4db8b1c6/standalone-metastore/src/main/sql/derby/hive-schema-4.0.0.derby.sql
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/hive/blob/4db8b1c6/standalone-metastore/src/main/sql/mssql/hive-schema-4.0.0.mssql.sql
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/hive/blob/4db8b1c6/standalone-metastore/src/main/sql/mysql/hive-schema-4.0.0.mysql.sql
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/hive/blob/4db8b1c6/standalone-metastore/src/main/sql/oracle/hive-schema-4.0.0.oracle.sql
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/hive/blob/4db8b1c6/standalone-metastore/src/main/sql/oracle/upgrade-2.3.0-to-3.0.0.oracle.sql
----------------------------------------------------------------------
diff --cc standalone-metastore/src/main/sql/oracle/upgrade-2.3.0-to-3.0.0.oracle.sql
index 5b767bc,71f5034..c9c6b30
--- a/standalone-metastore/src/main/sql/oracle/upgrade-2.3.0-to-3.0.0.oracle.sql
+++ b/standalone-metastore/src/main/sql/oracle/upgrade-2.3.0-to-3.0.0.oracle.sql
@@@ -338,5 -340,3 +340,4 @@@ ALTER TABLE TBLS ADD OWNER_TYPE VARCHAR
  -- These lines need to be last.  Insert any changes above.
  UPDATE VERSION SET SCHEMA_VERSION='3.0.0', VERSION_COMMENT='Hive release version 3.0.0' where VER_ID=1;
  SELECT 'Finished upgrading MetaStore schema from 2.3.0 to 3.0.0' AS Status from dual;
 +
- ALTER TABLE TBLS ADD OWNER_TYPE VARCHAR2(10) NULL;

http://git-wip-us.apache.org/repos/asf/hive/blob/4db8b1c6/standalone-metastore/src/main/sql/postgres/hive-schema-4.0.0.postgres.sql
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/hive/blob/4db8b1c6/standalone-metastore/src/main/thrift/hive_metastore.thrift
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/hive/blob/4db8b1c6/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/DummyRawStoreControlledCommit.java
----------------------------------------------------------------------
diff --cc standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/DummyRawStoreControlledCommit.java
index 70771cd,2454479..8729883
--- a/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/DummyRawStoreControlledCommit.java
+++ b/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/DummyRawStoreControlledCommit.java
@@@ -73,18 -85,7 +73,19 @@@ import org.apache.hadoop.hive.metastore
  import org.apache.hadoop.hive.metastore.api.UnknownTableException;
  import org.apache.hadoop.hive.metastore.api.WMMapping;
  import org.apache.hadoop.hive.metastore.api.WMPool;
 +import org.apache.hadoop.hive.metastore.api.WMFullResourcePlan;
 +import org.apache.hadoop.hive.metastore.api.ISchemaName;
 +import org.apache.hadoop.hive.metastore.api.SchemaVersionDescriptor;
 +
 +import java.nio.ByteBuffer;
 +import java.util.ArrayList;
 +import java.util.Collections;
 +import java.util.List;
 +import java.util.Map;
 +
 +import org.apache.hadoop.conf.Configurable;
 +import org.apache.hadoop.conf.Configuration;
+ import org.apache.hadoop.hive.metastore.api.WriteEventInfo;
  import org.apache.hadoop.hive.metastore.partition.spec.PartitionSpecProxy;
  import org.apache.hadoop.hive.metastore.utils.MetaStoreUtils.ColStatsObjWithSourceInfo;
  import org.apache.thrift.TException;

http://git-wip-us.apache.org/repos/asf/hive/blob/4db8b1c6/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/DummyRawStoreForJdoConnection.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/hive/blob/4db8b1c6/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClientPreCatalog.java
----------------------------------------------------------------------