You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by ha...@apache.org on 2014/01/16 23:01:54 UTC
svn commit: r1558928 [4/8] - in /hive/trunk: metastore/if/
metastore/src/gen/thrift/gen-cpp/
metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/
metastore/src/gen/thrift/gen-php/metastore/
metastore/src/gen/thrift/gen-py/hive_me...
Modified: hive/trunk/metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ThriftHiveMetastore.java
URL: http://svn.apache.org/viewvc/hive/trunk/metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ThriftHiveMetastore.java?rev=1558928&r1=1558927&r2=1558928&view=diff
==============================================================================
--- hive/trunk/metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ThriftHiveMetastore.java (original)
+++ hive/trunk/metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ThriftHiveMetastore.java Thu Jan 16 22:01:53 2014
@@ -92,6 +92,8 @@ public class ThriftHiveMetastore {
public Partition append_partition(String db_name, String tbl_name, List<String> part_vals) throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException;
+ public AddPartitionsResult add_partitions_req(AddPartitionsRequest request) throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException;
+
public Partition append_partition_with_environment_context(String db_name, String tbl_name, List<String> part_vals, EnvironmentContext environment_context) throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException;
public Partition append_partition_by_name(String db_name, String tbl_name, String part_name) throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException;
@@ -262,6 +264,8 @@ public class ThriftHiveMetastore {
public void append_partition(String db_name, String tbl_name, List<String> part_vals, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.append_partition_call> resultHandler) throws org.apache.thrift.TException;
+ public void add_partitions_req(AddPartitionsRequest request, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.add_partitions_req_call> resultHandler) throws org.apache.thrift.TException;
+
public void append_partition_with_environment_context(String db_name, String tbl_name, List<String> part_vals, EnvironmentContext environment_context, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.append_partition_with_environment_context_call> resultHandler) throws org.apache.thrift.TException;
public void append_partition_by_name(String db_name, String tbl_name, String part_name, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.append_partition_by_name_call> resultHandler) throws org.apache.thrift.TException;
@@ -1205,6 +1209,38 @@ public class ThriftHiveMetastore {
throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "append_partition failed: unknown result");
}
+ public AddPartitionsResult add_partitions_req(AddPartitionsRequest request) throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException
+ {
+ send_add_partitions_req(request);
+ return recv_add_partitions_req();
+ }
+
+ public void send_add_partitions_req(AddPartitionsRequest request) throws org.apache.thrift.TException
+ {
+ add_partitions_req_args args = new add_partitions_req_args();
+ args.setRequest(request);
+ sendBase("add_partitions_req", args);
+ }
+
+ public AddPartitionsResult recv_add_partitions_req() throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException
+ {
+ add_partitions_req_result result = new add_partitions_req_result();
+ receiveBase(result, "add_partitions_req");
+ if (result.isSetSuccess()) {
+ return result.success;
+ }
+ if (result.o1 != null) {
+ throw result.o1;
+ }
+ if (result.o2 != null) {
+ throw result.o2;
+ }
+ if (result.o3 != null) {
+ throw result.o3;
+ }
+ throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "add_partitions_req failed: unknown result");
+ }
+
public Partition append_partition_with_environment_context(String db_name, String tbl_name, List<String> part_vals, EnvironmentContext environment_context) throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException
{
send_append_partition_with_environment_context(db_name, tbl_name, part_vals, environment_context);
@@ -3889,6 +3925,38 @@ public class ThriftHiveMetastore {
}
}
+ public void add_partitions_req(AddPartitionsRequest request, org.apache.thrift.async.AsyncMethodCallback<add_partitions_req_call> resultHandler) throws org.apache.thrift.TException {
+ checkReady();
+ add_partitions_req_call method_call = new add_partitions_req_call(request, resultHandler, this, ___protocolFactory, ___transport);
+ this.___currentMethod = method_call;
+ ___manager.call(method_call);
+ }
+
+ public static class add_partitions_req_call extends org.apache.thrift.async.TAsyncMethodCall {
+ private AddPartitionsRequest request;
+ public add_partitions_req_call(AddPartitionsRequest request, org.apache.thrift.async.AsyncMethodCallback<add_partitions_req_call> resultHandler, org.apache.thrift.async.TAsyncClient client, org.apache.thrift.protocol.TProtocolFactory protocolFactory, org.apache.thrift.transport.TNonblockingTransport transport) throws org.apache.thrift.TException {
+ super(client, protocolFactory, transport, resultHandler, false);
+ this.request = request;
+ }
+
+ public void write_args(org.apache.thrift.protocol.TProtocol prot) throws org.apache.thrift.TException {
+ prot.writeMessageBegin(new org.apache.thrift.protocol.TMessage("add_partitions_req", org.apache.thrift.protocol.TMessageType.CALL, 0));
+ add_partitions_req_args args = new add_partitions_req_args();
+ args.setRequest(request);
+ args.write(prot);
+ prot.writeMessageEnd();
+ }
+
+ public AddPartitionsResult getResult() throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException {
+ if (getState() != org.apache.thrift.async.TAsyncMethodCall.State.RESPONSE_READ) {
+ throw new IllegalStateException("Method call not finished!");
+ }
+ org.apache.thrift.transport.TMemoryInputTransport memoryTransport = new org.apache.thrift.transport.TMemoryInputTransport(getFrameBuffer().array());
+ org.apache.thrift.protocol.TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport);
+ return (new Client(prot)).recv_add_partitions_req();
+ }
+ }
+
public void append_partition_with_environment_context(String db_name, String tbl_name, List<String> part_vals, EnvironmentContext environment_context, org.apache.thrift.async.AsyncMethodCallback<append_partition_with_environment_context_call> resultHandler) throws org.apache.thrift.TException {
checkReady();
append_partition_with_environment_context_call method_call = new append_partition_with_environment_context_call(db_name, tbl_name, part_vals, environment_context, resultHandler, this, ___protocolFactory, ___transport);
@@ -6057,6 +6125,7 @@ public class ThriftHiveMetastore {
processMap.put("add_partition_with_environment_context", new add_partition_with_environment_context());
processMap.put("add_partitions", new add_partitions());
processMap.put("append_partition", new append_partition());
+ processMap.put("add_partitions_req", new add_partitions_req());
processMap.put("append_partition_with_environment_context", new append_partition_with_environment_context());
processMap.put("append_partition_by_name", new append_partition_by_name());
processMap.put("append_partition_by_name_with_environment_context", new append_partition_by_name_with_environment_context());
@@ -6841,6 +6910,34 @@ public class ThriftHiveMetastore {
}
}
+ public static class add_partitions_req<I extends Iface> extends org.apache.thrift.ProcessFunction<I, add_partitions_req_args> {
+ public add_partitions_req() {
+ super("add_partitions_req");
+ }
+
+ public add_partitions_req_args getEmptyArgsInstance() {
+ return new add_partitions_req_args();
+ }
+
+ protected boolean isOneway() {
+ return false;
+ }
+
+ public add_partitions_req_result getResult(I iface, add_partitions_req_args args) throws org.apache.thrift.TException {
+ add_partitions_req_result result = new add_partitions_req_result();
+ try {
+ result.success = iface.add_partitions_req(args.request);
+ } catch (InvalidObjectException o1) {
+ result.o1 = o1;
+ } catch (AlreadyExistsException o2) {
+ result.o2 = o2;
+ } catch (MetaException o3) {
+ result.o3 = o3;
+ }
+ return result;
+ }
+ }
+
public static class append_partition_with_environment_context<I extends Iface> extends org.apache.thrift.ProcessFunction<I, append_partition_with_environment_context_args> {
public append_partition_with_environment_context() {
super("append_partition_with_environment_context");
@@ -12081,13 +12178,13 @@ public class ThriftHiveMetastore {
case 0: // SUCCESS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list264 = iprot.readListBegin();
- struct.success = new ArrayList<String>(_list264.size);
- for (int _i265 = 0; _i265 < _list264.size; ++_i265)
+ org.apache.thrift.protocol.TList _list280 = iprot.readListBegin();
+ struct.success = new ArrayList<String>(_list280.size);
+ for (int _i281 = 0; _i281 < _list280.size; ++_i281)
{
- String _elem266; // required
- _elem266 = iprot.readString();
- struct.success.add(_elem266);
+ String _elem282; // required
+ _elem282 = iprot.readString();
+ struct.success.add(_elem282);
}
iprot.readListEnd();
}
@@ -12122,9 +12219,9 @@ public class ThriftHiveMetastore {
oprot.writeFieldBegin(SUCCESS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size()));
- for (String _iter267 : struct.success)
+ for (String _iter283 : struct.success)
{
- oprot.writeString(_iter267);
+ oprot.writeString(_iter283);
}
oprot.writeListEnd();
}
@@ -12163,9 +12260,9 @@ public class ThriftHiveMetastore {
if (struct.isSetSuccess()) {
{
oprot.writeI32(struct.success.size());
- for (String _iter268 : struct.success)
+ for (String _iter284 : struct.success)
{
- oprot.writeString(_iter268);
+ oprot.writeString(_iter284);
}
}
}
@@ -12180,13 +12277,13 @@ public class ThriftHiveMetastore {
BitSet incoming = iprot.readBitSet(2);
if (incoming.get(0)) {
{
- org.apache.thrift.protocol.TList _list269 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
- struct.success = new ArrayList<String>(_list269.size);
- for (int _i270 = 0; _i270 < _list269.size; ++_i270)
+ org.apache.thrift.protocol.TList _list285 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
+ struct.success = new ArrayList<String>(_list285.size);
+ for (int _i286 = 0; _i286 < _list285.size; ++_i286)
{
- String _elem271; // required
- _elem271 = iprot.readString();
- struct.success.add(_elem271);
+ String _elem287; // required
+ _elem287 = iprot.readString();
+ struct.success.add(_elem287);
}
}
struct.setSuccessIsSet(true);
@@ -12843,13 +12940,13 @@ public class ThriftHiveMetastore {
case 0: // SUCCESS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list272 = iprot.readListBegin();
- struct.success = new ArrayList<String>(_list272.size);
- for (int _i273 = 0; _i273 < _list272.size; ++_i273)
+ org.apache.thrift.protocol.TList _list288 = iprot.readListBegin();
+ struct.success = new ArrayList<String>(_list288.size);
+ for (int _i289 = 0; _i289 < _list288.size; ++_i289)
{
- String _elem274; // required
- _elem274 = iprot.readString();
- struct.success.add(_elem274);
+ String _elem290; // required
+ _elem290 = iprot.readString();
+ struct.success.add(_elem290);
}
iprot.readListEnd();
}
@@ -12884,9 +12981,9 @@ public class ThriftHiveMetastore {
oprot.writeFieldBegin(SUCCESS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size()));
- for (String _iter275 : struct.success)
+ for (String _iter291 : struct.success)
{
- oprot.writeString(_iter275);
+ oprot.writeString(_iter291);
}
oprot.writeListEnd();
}
@@ -12925,9 +13022,9 @@ public class ThriftHiveMetastore {
if (struct.isSetSuccess()) {
{
oprot.writeI32(struct.success.size());
- for (String _iter276 : struct.success)
+ for (String _iter292 : struct.success)
{
- oprot.writeString(_iter276);
+ oprot.writeString(_iter292);
}
}
}
@@ -12942,13 +13039,13 @@ public class ThriftHiveMetastore {
BitSet incoming = iprot.readBitSet(2);
if (incoming.get(0)) {
{
- org.apache.thrift.protocol.TList _list277 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
- struct.success = new ArrayList<String>(_list277.size);
- for (int _i278 = 0; _i278 < _list277.size; ++_i278)
+ org.apache.thrift.protocol.TList _list293 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
+ struct.success = new ArrayList<String>(_list293.size);
+ for (int _i294 = 0; _i294 < _list293.size; ++_i294)
{
- String _elem279; // required
- _elem279 = iprot.readString();
- struct.success.add(_elem279);
+ String _elem295; // required
+ _elem295 = iprot.readString();
+ struct.success.add(_elem295);
}
}
struct.setSuccessIsSet(true);
@@ -17555,16 +17652,16 @@ public class ThriftHiveMetastore {
case 0: // SUCCESS
if (schemeField.type == org.apache.thrift.protocol.TType.MAP) {
{
- org.apache.thrift.protocol.TMap _map280 = iprot.readMapBegin();
- struct.success = new HashMap<String,Type>(2*_map280.size);
- for (int _i281 = 0; _i281 < _map280.size; ++_i281)
+ org.apache.thrift.protocol.TMap _map296 = iprot.readMapBegin();
+ struct.success = new HashMap<String,Type>(2*_map296.size);
+ for (int _i297 = 0; _i297 < _map296.size; ++_i297)
{
- String _key282; // required
- Type _val283; // required
- _key282 = iprot.readString();
- _val283 = new Type();
- _val283.read(iprot);
- struct.success.put(_key282, _val283);
+ String _key298; // required
+ Type _val299; // required
+ _key298 = iprot.readString();
+ _val299 = new Type();
+ _val299.read(iprot);
+ struct.success.put(_key298, _val299);
}
iprot.readMapEnd();
}
@@ -17599,10 +17696,10 @@ public class ThriftHiveMetastore {
oprot.writeFieldBegin(SUCCESS_FIELD_DESC);
{
oprot.writeMapBegin(new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRUCT, struct.success.size()));
- for (Map.Entry<String, Type> _iter284 : struct.success.entrySet())
+ for (Map.Entry<String, Type> _iter300 : struct.success.entrySet())
{
- oprot.writeString(_iter284.getKey());
- _iter284.getValue().write(oprot);
+ oprot.writeString(_iter300.getKey());
+ _iter300.getValue().write(oprot);
}
oprot.writeMapEnd();
}
@@ -17641,10 +17738,10 @@ public class ThriftHiveMetastore {
if (struct.isSetSuccess()) {
{
oprot.writeI32(struct.success.size());
- for (Map.Entry<String, Type> _iter285 : struct.success.entrySet())
+ for (Map.Entry<String, Type> _iter301 : struct.success.entrySet())
{
- oprot.writeString(_iter285.getKey());
- _iter285.getValue().write(oprot);
+ oprot.writeString(_iter301.getKey());
+ _iter301.getValue().write(oprot);
}
}
}
@@ -17659,16 +17756,16 @@ public class ThriftHiveMetastore {
BitSet incoming = iprot.readBitSet(2);
if (incoming.get(0)) {
{
- org.apache.thrift.protocol.TMap _map286 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
- struct.success = new HashMap<String,Type>(2*_map286.size);
- for (int _i287 = 0; _i287 < _map286.size; ++_i287)
+ org.apache.thrift.protocol.TMap _map302 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
+ struct.success = new HashMap<String,Type>(2*_map302.size);
+ for (int _i303 = 0; _i303 < _map302.size; ++_i303)
{
- String _key288; // required
- Type _val289; // required
- _key288 = iprot.readString();
- _val289 = new Type();
- _val289.read(iprot);
- struct.success.put(_key288, _val289);
+ String _key304; // required
+ Type _val305; // required
+ _key304 = iprot.readString();
+ _val305 = new Type();
+ _val305.read(iprot);
+ struct.success.put(_key304, _val305);
}
}
struct.setSuccessIsSet(true);
@@ -18703,14 +18800,14 @@ public class ThriftHiveMetastore {
case 0: // SUCCESS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list290 = iprot.readListBegin();
- struct.success = new ArrayList<FieldSchema>(_list290.size);
- for (int _i291 = 0; _i291 < _list290.size; ++_i291)
+ org.apache.thrift.protocol.TList _list306 = iprot.readListBegin();
+ struct.success = new ArrayList<FieldSchema>(_list306.size);
+ for (int _i307 = 0; _i307 < _list306.size; ++_i307)
{
- FieldSchema _elem292; // required
- _elem292 = new FieldSchema();
- _elem292.read(iprot);
- struct.success.add(_elem292);
+ FieldSchema _elem308; // required
+ _elem308 = new FieldSchema();
+ _elem308.read(iprot);
+ struct.success.add(_elem308);
}
iprot.readListEnd();
}
@@ -18763,9 +18860,9 @@ public class ThriftHiveMetastore {
oprot.writeFieldBegin(SUCCESS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size()));
- for (FieldSchema _iter293 : struct.success)
+ for (FieldSchema _iter309 : struct.success)
{
- _iter293.write(oprot);
+ _iter309.write(oprot);
}
oprot.writeListEnd();
}
@@ -18820,9 +18917,9 @@ public class ThriftHiveMetastore {
if (struct.isSetSuccess()) {
{
oprot.writeI32(struct.success.size());
- for (FieldSchema _iter294 : struct.success)
+ for (FieldSchema _iter310 : struct.success)
{
- _iter294.write(oprot);
+ _iter310.write(oprot);
}
}
}
@@ -18843,14 +18940,14 @@ public class ThriftHiveMetastore {
BitSet incoming = iprot.readBitSet(4);
if (incoming.get(0)) {
{
- org.apache.thrift.protocol.TList _list295 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
- struct.success = new ArrayList<FieldSchema>(_list295.size);
- for (int _i296 = 0; _i296 < _list295.size; ++_i296)
+ org.apache.thrift.protocol.TList _list311 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
+ struct.success = new ArrayList<FieldSchema>(_list311.size);
+ for (int _i312 = 0; _i312 < _list311.size; ++_i312)
{
- FieldSchema _elem297; // required
- _elem297 = new FieldSchema();
- _elem297.read(iprot);
- struct.success.add(_elem297);
+ FieldSchema _elem313; // required
+ _elem313 = new FieldSchema();
+ _elem313.read(iprot);
+ struct.success.add(_elem313);
}
}
struct.setSuccessIsSet(true);
@@ -19895,14 +19992,14 @@ public class ThriftHiveMetastore {
case 0: // SUCCESS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list298 = iprot.readListBegin();
- struct.success = new ArrayList<FieldSchema>(_list298.size);
- for (int _i299 = 0; _i299 < _list298.size; ++_i299)
+ org.apache.thrift.protocol.TList _list314 = iprot.readListBegin();
+ struct.success = new ArrayList<FieldSchema>(_list314.size);
+ for (int _i315 = 0; _i315 < _list314.size; ++_i315)
{
- FieldSchema _elem300; // required
- _elem300 = new FieldSchema();
- _elem300.read(iprot);
- struct.success.add(_elem300);
+ FieldSchema _elem316; // required
+ _elem316 = new FieldSchema();
+ _elem316.read(iprot);
+ struct.success.add(_elem316);
}
iprot.readListEnd();
}
@@ -19955,9 +20052,9 @@ public class ThriftHiveMetastore {
oprot.writeFieldBegin(SUCCESS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size()));
- for (FieldSchema _iter301 : struct.success)
+ for (FieldSchema _iter317 : struct.success)
{
- _iter301.write(oprot);
+ _iter317.write(oprot);
}
oprot.writeListEnd();
}
@@ -20012,9 +20109,9 @@ public class ThriftHiveMetastore {
if (struct.isSetSuccess()) {
{
oprot.writeI32(struct.success.size());
- for (FieldSchema _iter302 : struct.success)
+ for (FieldSchema _iter318 : struct.success)
{
- _iter302.write(oprot);
+ _iter318.write(oprot);
}
}
}
@@ -20035,14 +20132,14 @@ public class ThriftHiveMetastore {
BitSet incoming = iprot.readBitSet(4);
if (incoming.get(0)) {
{
- org.apache.thrift.protocol.TList _list303 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
- struct.success = new ArrayList<FieldSchema>(_list303.size);
- for (int _i304 = 0; _i304 < _list303.size; ++_i304)
+ org.apache.thrift.protocol.TList _list319 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
+ struct.success = new ArrayList<FieldSchema>(_list319.size);
+ for (int _i320 = 0; _i320 < _list319.size; ++_i320)
{
- FieldSchema _elem305; // required
- _elem305 = new FieldSchema();
- _elem305.read(iprot);
- struct.success.add(_elem305);
+ FieldSchema _elem321; // required
+ _elem321 = new FieldSchema();
+ _elem321.read(iprot);
+ struct.success.add(_elem321);
}
}
struct.setSuccessIsSet(true);
@@ -25285,13 +25382,13 @@ public class ThriftHiveMetastore {
case 0: // SUCCESS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list306 = iprot.readListBegin();
- struct.success = new ArrayList<String>(_list306.size);
- for (int _i307 = 0; _i307 < _list306.size; ++_i307)
+ org.apache.thrift.protocol.TList _list322 = iprot.readListBegin();
+ struct.success = new ArrayList<String>(_list322.size);
+ for (int _i323 = 0; _i323 < _list322.size; ++_i323)
{
- String _elem308; // required
- _elem308 = iprot.readString();
- struct.success.add(_elem308);
+ String _elem324; // required
+ _elem324 = iprot.readString();
+ struct.success.add(_elem324);
}
iprot.readListEnd();
}
@@ -25326,9 +25423,9 @@ public class ThriftHiveMetastore {
oprot.writeFieldBegin(SUCCESS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size()));
- for (String _iter309 : struct.success)
+ for (String _iter325 : struct.success)
{
- oprot.writeString(_iter309);
+ oprot.writeString(_iter325);
}
oprot.writeListEnd();
}
@@ -25367,9 +25464,9 @@ public class ThriftHiveMetastore {
if (struct.isSetSuccess()) {
{
oprot.writeI32(struct.success.size());
- for (String _iter310 : struct.success)
+ for (String _iter326 : struct.success)
{
- oprot.writeString(_iter310);
+ oprot.writeString(_iter326);
}
}
}
@@ -25384,13 +25481,13 @@ public class ThriftHiveMetastore {
BitSet incoming = iprot.readBitSet(2);
if (incoming.get(0)) {
{
- org.apache.thrift.protocol.TList _list311 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
- struct.success = new ArrayList<String>(_list311.size);
- for (int _i312 = 0; _i312 < _list311.size; ++_i312)
+ org.apache.thrift.protocol.TList _list327 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
+ struct.success = new ArrayList<String>(_list327.size);
+ for (int _i328 = 0; _i328 < _list327.size; ++_i328)
{
- String _elem313; // required
- _elem313 = iprot.readString();
- struct.success.add(_elem313);
+ String _elem329; // required
+ _elem329 = iprot.readString();
+ struct.success.add(_elem329);
}
}
struct.setSuccessIsSet(true);
@@ -26159,13 +26256,13 @@ public class ThriftHiveMetastore {
case 0: // SUCCESS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list314 = iprot.readListBegin();
- struct.success = new ArrayList<String>(_list314.size);
- for (int _i315 = 0; _i315 < _list314.size; ++_i315)
+ org.apache.thrift.protocol.TList _list330 = iprot.readListBegin();
+ struct.success = new ArrayList<String>(_list330.size);
+ for (int _i331 = 0; _i331 < _list330.size; ++_i331)
{
- String _elem316; // required
- _elem316 = iprot.readString();
- struct.success.add(_elem316);
+ String _elem332; // required
+ _elem332 = iprot.readString();
+ struct.success.add(_elem332);
}
iprot.readListEnd();
}
@@ -26200,9 +26297,9 @@ public class ThriftHiveMetastore {
oprot.writeFieldBegin(SUCCESS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size()));
- for (String _iter317 : struct.success)
+ for (String _iter333 : struct.success)
{
- oprot.writeString(_iter317);
+ oprot.writeString(_iter333);
}
oprot.writeListEnd();
}
@@ -26241,9 +26338,9 @@ public class ThriftHiveMetastore {
if (struct.isSetSuccess()) {
{
oprot.writeI32(struct.success.size());
- for (String _iter318 : struct.success)
+ for (String _iter334 : struct.success)
{
- oprot.writeString(_iter318);
+ oprot.writeString(_iter334);
}
}
}
@@ -26258,13 +26355,13 @@ public class ThriftHiveMetastore {
BitSet incoming = iprot.readBitSet(2);
if (incoming.get(0)) {
{
- org.apache.thrift.protocol.TList _list319 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
- struct.success = new ArrayList<String>(_list319.size);
- for (int _i320 = 0; _i320 < _list319.size; ++_i320)
+ org.apache.thrift.protocol.TList _list335 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
+ struct.success = new ArrayList<String>(_list335.size);
+ for (int _i336 = 0; _i336 < _list335.size; ++_i336)
{
- String _elem321; // required
- _elem321 = iprot.readString();
- struct.success.add(_elem321);
+ String _elem337; // required
+ _elem337 = iprot.readString();
+ struct.success.add(_elem337);
}
}
struct.setSuccessIsSet(true);
@@ -27720,13 +27817,13 @@ public class ThriftHiveMetastore {
case 2: // TBL_NAMES
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list322 = iprot.readListBegin();
- struct.tbl_names = new ArrayList<String>(_list322.size);
- for (int _i323 = 0; _i323 < _list322.size; ++_i323)
+ org.apache.thrift.protocol.TList _list338 = iprot.readListBegin();
+ struct.tbl_names = new ArrayList<String>(_list338.size);
+ for (int _i339 = 0; _i339 < _list338.size; ++_i339)
{
- String _elem324; // required
- _elem324 = iprot.readString();
- struct.tbl_names.add(_elem324);
+ String _elem340; // required
+ _elem340 = iprot.readString();
+ struct.tbl_names.add(_elem340);
}
iprot.readListEnd();
}
@@ -27757,9 +27854,9 @@ public class ThriftHiveMetastore {
oprot.writeFieldBegin(TBL_NAMES_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.tbl_names.size()));
- for (String _iter325 : struct.tbl_names)
+ for (String _iter341 : struct.tbl_names)
{
- oprot.writeString(_iter325);
+ oprot.writeString(_iter341);
}
oprot.writeListEnd();
}
@@ -27796,9 +27893,9 @@ public class ThriftHiveMetastore {
if (struct.isSetTbl_names()) {
{
oprot.writeI32(struct.tbl_names.size());
- for (String _iter326 : struct.tbl_names)
+ for (String _iter342 : struct.tbl_names)
{
- oprot.writeString(_iter326);
+ oprot.writeString(_iter342);
}
}
}
@@ -27814,13 +27911,13 @@ public class ThriftHiveMetastore {
}
if (incoming.get(1)) {
{
- org.apache.thrift.protocol.TList _list327 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
- struct.tbl_names = new ArrayList<String>(_list327.size);
- for (int _i328 = 0; _i328 < _list327.size; ++_i328)
+ org.apache.thrift.protocol.TList _list343 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
+ struct.tbl_names = new ArrayList<String>(_list343.size);
+ for (int _i344 = 0; _i344 < _list343.size; ++_i344)
{
- String _elem329; // required
- _elem329 = iprot.readString();
- struct.tbl_names.add(_elem329);
+ String _elem345; // required
+ _elem345 = iprot.readString();
+ struct.tbl_names.add(_elem345);
}
}
struct.setTbl_namesIsSet(true);
@@ -28388,14 +28485,14 @@ public class ThriftHiveMetastore {
case 0: // SUCCESS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list330 = iprot.readListBegin();
- struct.success = new ArrayList<Table>(_list330.size);
- for (int _i331 = 0; _i331 < _list330.size; ++_i331)
+ org.apache.thrift.protocol.TList _list346 = iprot.readListBegin();
+ struct.success = new ArrayList<Table>(_list346.size);
+ for (int _i347 = 0; _i347 < _list346.size; ++_i347)
{
- Table _elem332; // required
- _elem332 = new Table();
- _elem332.read(iprot);
- struct.success.add(_elem332);
+ Table _elem348; // required
+ _elem348 = new Table();
+ _elem348.read(iprot);
+ struct.success.add(_elem348);
}
iprot.readListEnd();
}
@@ -28448,9 +28545,9 @@ public class ThriftHiveMetastore {
oprot.writeFieldBegin(SUCCESS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size()));
- for (Table _iter333 : struct.success)
+ for (Table _iter349 : struct.success)
{
- _iter333.write(oprot);
+ _iter349.write(oprot);
}
oprot.writeListEnd();
}
@@ -28505,9 +28602,9 @@ public class ThriftHiveMetastore {
if (struct.isSetSuccess()) {
{
oprot.writeI32(struct.success.size());
- for (Table _iter334 : struct.success)
+ for (Table _iter350 : struct.success)
{
- _iter334.write(oprot);
+ _iter350.write(oprot);
}
}
}
@@ -28528,14 +28625,14 @@ public class ThriftHiveMetastore {
BitSet incoming = iprot.readBitSet(4);
if (incoming.get(0)) {
{
- org.apache.thrift.protocol.TList _list335 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
- struct.success = new ArrayList<Table>(_list335.size);
- for (int _i336 = 0; _i336 < _list335.size; ++_i336)
+ org.apache.thrift.protocol.TList _list351 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
+ struct.success = new ArrayList<Table>(_list351.size);
+ for (int _i352 = 0; _i352 < _list351.size; ++_i352)
{
- Table _elem337; // required
- _elem337 = new Table();
- _elem337.read(iprot);
- struct.success.add(_elem337);
+ Table _elem353; // required
+ _elem353 = new Table();
+ _elem353.read(iprot);
+ struct.success.add(_elem353);
}
}
struct.setSuccessIsSet(true);
@@ -29684,13 +29781,13 @@ public class ThriftHiveMetastore {
case 0: // SUCCESS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list338 = iprot.readListBegin();
- struct.success = new ArrayList<String>(_list338.size);
- for (int _i339 = 0; _i339 < _list338.size; ++_i339)
+ org.apache.thrift.protocol.TList _list354 = iprot.readListBegin();
+ struct.success = new ArrayList<String>(_list354.size);
+ for (int _i355 = 0; _i355 < _list354.size; ++_i355)
{
- String _elem340; // required
- _elem340 = iprot.readString();
- struct.success.add(_elem340);
+ String _elem356; // required
+ _elem356 = iprot.readString();
+ struct.success.add(_elem356);
}
iprot.readListEnd();
}
@@ -29743,9 +29840,9 @@ public class ThriftHiveMetastore {
oprot.writeFieldBegin(SUCCESS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size()));
- for (String _iter341 : struct.success)
+ for (String _iter357 : struct.success)
{
- oprot.writeString(_iter341);
+ oprot.writeString(_iter357);
}
oprot.writeListEnd();
}
@@ -29800,9 +29897,9 @@ public class ThriftHiveMetastore {
if (struct.isSetSuccess()) {
{
oprot.writeI32(struct.success.size());
- for (String _iter342 : struct.success)
+ for (String _iter358 : struct.success)
{
- oprot.writeString(_iter342);
+ oprot.writeString(_iter358);
}
}
}
@@ -29823,13 +29920,13 @@ public class ThriftHiveMetastore {
BitSet incoming = iprot.readBitSet(4);
if (incoming.get(0)) {
{
- org.apache.thrift.protocol.TList _list343 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
- struct.success = new ArrayList<String>(_list343.size);
- for (int _i344 = 0; _i344 < _list343.size; ++_i344)
+ org.apache.thrift.protocol.TList _list359 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
+ struct.success = new ArrayList<String>(_list359.size);
+ for (int _i360 = 0; _i360 < _list359.size; ++_i360)
{
- String _elem345; // required
- _elem345 = iprot.readString();
- struct.success.add(_elem345);
+ String _elem361; // required
+ _elem361 = iprot.readString();
+ struct.success.add(_elem361);
}
}
struct.setSuccessIsSet(true);
@@ -34549,14 +34646,14 @@ public class ThriftHiveMetastore {
case 1: // NEW_PARTS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list346 = iprot.readListBegin();
- struct.new_parts = new ArrayList<Partition>(_list346.size);
- for (int _i347 = 0; _i347 < _list346.size; ++_i347)
+ org.apache.thrift.protocol.TList _list362 = iprot.readListBegin();
+ struct.new_parts = new ArrayList<Partition>(_list362.size);
+ for (int _i363 = 0; _i363 < _list362.size; ++_i363)
{
- Partition _elem348; // required
- _elem348 = new Partition();
- _elem348.read(iprot);
- struct.new_parts.add(_elem348);
+ Partition _elem364; // required
+ _elem364 = new Partition();
+ _elem364.read(iprot);
+ struct.new_parts.add(_elem364);
}
iprot.readListEnd();
}
@@ -34582,9 +34679,9 @@ public class ThriftHiveMetastore {
oprot.writeFieldBegin(NEW_PARTS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.new_parts.size()));
- for (Partition _iter349 : struct.new_parts)
+ for (Partition _iter365 : struct.new_parts)
{
- _iter349.write(oprot);
+ _iter365.write(oprot);
}
oprot.writeListEnd();
}
@@ -34615,9 +34712,9 @@ public class ThriftHiveMetastore {
if (struct.isSetNew_parts()) {
{
oprot.writeI32(struct.new_parts.size());
- for (Partition _iter350 : struct.new_parts)
+ for (Partition _iter366 : struct.new_parts)
{
- _iter350.write(oprot);
+ _iter366.write(oprot);
}
}
}
@@ -34629,14 +34726,14 @@ public class ThriftHiveMetastore {
BitSet incoming = iprot.readBitSet(1);
if (incoming.get(0)) {
{
- org.apache.thrift.protocol.TList _list351 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
- struct.new_parts = new ArrayList<Partition>(_list351.size);
- for (int _i352 = 0; _i352 < _list351.size; ++_i352)
+ org.apache.thrift.protocol.TList _list367 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
+ struct.new_parts = new ArrayList<Partition>(_list367.size);
+ for (int _i368 = 0; _i368 < _list367.size; ++_i368)
{
- Partition _elem353; // required
- _elem353 = new Partition();
- _elem353.read(iprot);
- struct.new_parts.add(_elem353);
+ Partition _elem369; // required
+ _elem369 = new Partition();
+ _elem369.read(iprot);
+ struct.new_parts.add(_elem369);
}
}
struct.setNew_partsIsSet(true);
@@ -35815,13 +35912,13 @@ public class ThriftHiveMetastore {
case 3: // PART_VALS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list354 = iprot.readListBegin();
- struct.part_vals = new ArrayList<String>(_list354.size);
- for (int _i355 = 0; _i355 < _list354.size; ++_i355)
+ org.apache.thrift.protocol.TList _list370 = iprot.readListBegin();
+ struct.part_vals = new ArrayList<String>(_list370.size);
+ for (int _i371 = 0; _i371 < _list370.size; ++_i371)
{
- String _elem356; // required
- _elem356 = iprot.readString();
- struct.part_vals.add(_elem356);
+ String _elem372; // required
+ _elem372 = iprot.readString();
+ struct.part_vals.add(_elem372);
}
iprot.readListEnd();
}
@@ -35843,26 +35940,1104 @@ public class ThriftHiveMetastore {
struct.validate();
oprot.writeStructBegin(STRUCT_DESC);
- if (struct.db_name != null) {
- oprot.writeFieldBegin(DB_NAME_FIELD_DESC);
- oprot.writeString(struct.db_name);
- oprot.writeFieldEnd();
- }
- if (struct.tbl_name != null) {
- oprot.writeFieldBegin(TBL_NAME_FIELD_DESC);
- oprot.writeString(struct.tbl_name);
- oprot.writeFieldEnd();
- }
- if (struct.part_vals != null) {
- oprot.writeFieldBegin(PART_VALS_FIELD_DESC);
- {
- oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.part_vals.size()));
- for (String _iter357 : struct.part_vals)
- {
- oprot.writeString(_iter357);
- }
- oprot.writeListEnd();
- }
+ if (struct.db_name != null) {
+ oprot.writeFieldBegin(DB_NAME_FIELD_DESC);
+ oprot.writeString(struct.db_name);
+ oprot.writeFieldEnd();
+ }
+ if (struct.tbl_name != null) {
+ oprot.writeFieldBegin(TBL_NAME_FIELD_DESC);
+ oprot.writeString(struct.tbl_name);
+ oprot.writeFieldEnd();
+ }
+ if (struct.part_vals != null) {
+ oprot.writeFieldBegin(PART_VALS_FIELD_DESC);
+ {
+ oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.part_vals.size()));
+ for (String _iter373 : struct.part_vals)
+ {
+ oprot.writeString(_iter373);
+ }
+ oprot.writeListEnd();
+ }
+ oprot.writeFieldEnd();
+ }
+ oprot.writeFieldStop();
+ oprot.writeStructEnd();
+ }
+
+ }
+
+ private static class append_partition_argsTupleSchemeFactory implements SchemeFactory {
+ public append_partition_argsTupleScheme getScheme() {
+ return new append_partition_argsTupleScheme();
+ }
+ }
+
+ private static class append_partition_argsTupleScheme extends TupleScheme<append_partition_args> {
+
+ @Override
+ public void write(org.apache.thrift.protocol.TProtocol prot, append_partition_args struct) throws org.apache.thrift.TException {
+ TTupleProtocol oprot = (TTupleProtocol) prot;
+ BitSet optionals = new BitSet();
+ if (struct.isSetDb_name()) {
+ optionals.set(0);
+ }
+ if (struct.isSetTbl_name()) {
+ optionals.set(1);
+ }
+ if (struct.isSetPart_vals()) {
+ optionals.set(2);
+ }
+ oprot.writeBitSet(optionals, 3);
+ if (struct.isSetDb_name()) {
+ oprot.writeString(struct.db_name);
+ }
+ if (struct.isSetTbl_name()) {
+ oprot.writeString(struct.tbl_name);
+ }
+ if (struct.isSetPart_vals()) {
+ {
+ oprot.writeI32(struct.part_vals.size());
+ for (String _iter374 : struct.part_vals)
+ {
+ oprot.writeString(_iter374);
+ }
+ }
+ }
+ }
+
+ @Override
+ public void read(org.apache.thrift.protocol.TProtocol prot, append_partition_args struct) throws org.apache.thrift.TException {
+ TTupleProtocol iprot = (TTupleProtocol) prot;
+ BitSet incoming = iprot.readBitSet(3);
+ if (incoming.get(0)) {
+ struct.db_name = iprot.readString();
+ struct.setDb_nameIsSet(true);
+ }
+ if (incoming.get(1)) {
+ struct.tbl_name = iprot.readString();
+ struct.setTbl_nameIsSet(true);
+ }
+ if (incoming.get(2)) {
+ {
+ org.apache.thrift.protocol.TList _list375 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
+ struct.part_vals = new ArrayList<String>(_list375.size);
+ for (int _i376 = 0; _i376 < _list375.size; ++_i376)
+ {
+ String _elem377; // required
+ _elem377 = iprot.readString();
+ struct.part_vals.add(_elem377);
+ }
+ }
+ struct.setPart_valsIsSet(true);
+ }
+ }
+ }
+
+ }
+
+ public static class append_partition_result implements org.apache.thrift.TBase<append_partition_result, append_partition_result._Fields>, java.io.Serializable, Cloneable {
+ private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("append_partition_result");
+
+ private static final org.apache.thrift.protocol.TField SUCCESS_FIELD_DESC = new org.apache.thrift.protocol.TField("success", org.apache.thrift.protocol.TType.STRUCT, (short)0);
+ private static final org.apache.thrift.protocol.TField O1_FIELD_DESC = new org.apache.thrift.protocol.TField("o1", org.apache.thrift.protocol.TType.STRUCT, (short)1);
+ private static final org.apache.thrift.protocol.TField O2_FIELD_DESC = new org.apache.thrift.protocol.TField("o2", org.apache.thrift.protocol.TType.STRUCT, (short)2);
+ private static final org.apache.thrift.protocol.TField O3_FIELD_DESC = new org.apache.thrift.protocol.TField("o3", org.apache.thrift.protocol.TType.STRUCT, (short)3);
+
+ private static final Map<Class<? extends IScheme>, SchemeFactory> schemes = new HashMap<Class<? extends IScheme>, SchemeFactory>();
+ static {
+ schemes.put(StandardScheme.class, new append_partition_resultStandardSchemeFactory());
+ schemes.put(TupleScheme.class, new append_partition_resultTupleSchemeFactory());
+ }
+
+ private Partition success; // required
+ private InvalidObjectException o1; // required
+ private AlreadyExistsException o2; // required
+ private MetaException o3; // required
+
+ /** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */
+ public enum _Fields implements org.apache.thrift.TFieldIdEnum {
+ SUCCESS((short)0, "success"),
+ O1((short)1, "o1"),
+ O2((short)2, "o2"),
+ O3((short)3, "o3");
+
+ private static final Map<String, _Fields> byName = new HashMap<String, _Fields>();
+
+ static {
+ for (_Fields field : EnumSet.allOf(_Fields.class)) {
+ byName.put(field.getFieldName(), field);
+ }
+ }
+
+ /**
+ * Find the _Fields constant that matches fieldId, or null if its not found.
+ */
+ public static _Fields findByThriftId(int fieldId) {
+ switch(fieldId) {
+ case 0: // SUCCESS
+ return SUCCESS;
+ case 1: // O1
+ return O1;
+ case 2: // O2
+ return O2;
+ case 3: // O3
+ return O3;
+ default:
+ return null;
+ }
+ }
+
+ /**
+ * Find the _Fields constant that matches fieldId, throwing an exception
+ * if it is not found.
+ */
+ public static _Fields findByThriftIdOrThrow(int fieldId) {
+ _Fields fields = findByThriftId(fieldId);
+ if (fields == null) throw new IllegalArgumentException("Field " + fieldId + " doesn't exist!");
+ return fields;
+ }
+
+ /**
+ * Find the _Fields constant that matches name, or null if its not found.
+ */
+ public static _Fields findByName(String name) {
+ return byName.get(name);
+ }
+
+ private final short _thriftId;
+ private final String _fieldName;
+
+ _Fields(short thriftId, String fieldName) {
+ _thriftId = thriftId;
+ _fieldName = fieldName;
+ }
+
+ public short getThriftFieldId() {
+ return _thriftId;
+ }
+
+ public String getFieldName() {
+ return _fieldName;
+ }
+ }
+
+ // isset id assignments
+ public static final Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap;
+ static {
+ Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class);
+ tmpMap.put(_Fields.SUCCESS, new org.apache.thrift.meta_data.FieldMetaData("success", org.apache.thrift.TFieldRequirementType.DEFAULT,
+ new org.apache.thrift.meta_data.StructMetaData(org.apache.thrift.protocol.TType.STRUCT, Partition.class)));
+ tmpMap.put(_Fields.O1, new org.apache.thrift.meta_data.FieldMetaData("o1", org.apache.thrift.TFieldRequirementType.DEFAULT,
+ new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRUCT)));
+ tmpMap.put(_Fields.O2, new org.apache.thrift.meta_data.FieldMetaData("o2", org.apache.thrift.TFieldRequirementType.DEFAULT,
+ new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRUCT)));
+ tmpMap.put(_Fields.O3, new org.apache.thrift.meta_data.FieldMetaData("o3", org.apache.thrift.TFieldRequirementType.DEFAULT,
+ new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRUCT)));
+ metaDataMap = Collections.unmodifiableMap(tmpMap);
+ org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(append_partition_result.class, metaDataMap);
+ }
+
+ public append_partition_result() {
+ }
+
+ public append_partition_result(
+ Partition success,
+ InvalidObjectException o1,
+ AlreadyExistsException o2,
+ MetaException o3)
+ {
+ this();
+ this.success = success;
+ this.o1 = o1;
+ this.o2 = o2;
+ this.o3 = o3;
+ }
+
+ /**
+ * Performs a deep copy on <i>other</i>.
+ */
+ public append_partition_result(append_partition_result other) {
+ if (other.isSetSuccess()) {
+ this.success = new Partition(other.success);
+ }
+ if (other.isSetO1()) {
+ this.o1 = new InvalidObjectException(other.o1);
+ }
+ if (other.isSetO2()) {
+ this.o2 = new AlreadyExistsException(other.o2);
+ }
+ if (other.isSetO3()) {
+ this.o3 = new MetaException(other.o3);
+ }
+ }
+
+ public append_partition_result deepCopy() {
+ return new append_partition_result(this);
+ }
+
+ @Override
+ public void clear() {
+ this.success = null;
+ this.o1 = null;
+ this.o2 = null;
+ this.o3 = null;
+ }
+
+ public Partition getSuccess() {
+ return this.success;
+ }
+
+ public void setSuccess(Partition success) {
+ this.success = success;
+ }
+
+ public void unsetSuccess() {
+ this.success = null;
+ }
+
+ /** Returns true if field success is set (has been assigned a value) and false otherwise */
+ public boolean isSetSuccess() {
+ return this.success != null;
+ }
+
+ public void setSuccessIsSet(boolean value) {
+ if (!value) {
+ this.success = null;
+ }
+ }
+
+ public InvalidObjectException getO1() {
+ return this.o1;
+ }
+
+ public void setO1(InvalidObjectException o1) {
+ this.o1 = o1;
+ }
+
+ public void unsetO1() {
+ this.o1 = null;
+ }
+
+ /** Returns true if field o1 is set (has been assigned a value) and false otherwise */
+ public boolean isSetO1() {
+ return this.o1 != null;
+ }
+
+ public void setO1IsSet(boolean value) {
+ if (!value) {
+ this.o1 = null;
+ }
+ }
+
+ public AlreadyExistsException getO2() {
+ return this.o2;
+ }
+
+ public void setO2(AlreadyExistsException o2) {
+ this.o2 = o2;
+ }
+
+ public void unsetO2() {
+ this.o2 = null;
+ }
+
+ /** Returns true if field o2 is set (has been assigned a value) and false otherwise */
+ public boolean isSetO2() {
+ return this.o2 != null;
+ }
+
+ public void setO2IsSet(boolean value) {
+ if (!value) {
+ this.o2 = null;
+ }
+ }
+
+ public MetaException getO3() {
+ return this.o3;
+ }
+
+ public void setO3(MetaException o3) {
+ this.o3 = o3;
+ }
+
+ public void unsetO3() {
+ this.o3 = null;
+ }
+
+ /** Returns true if field o3 is set (has been assigned a value) and false otherwise */
+ public boolean isSetO3() {
+ return this.o3 != null;
+ }
+
+ public void setO3IsSet(boolean value) {
+ if (!value) {
+ this.o3 = null;
+ }
+ }
+
+ public void setFieldValue(_Fields field, Object value) {
+ switch (field) {
+ case SUCCESS:
+ if (value == null) {
+ unsetSuccess();
+ } else {
+ setSuccess((Partition)value);
+ }
+ break;
+
+ case O1:
+ if (value == null) {
+ unsetO1();
+ } else {
+ setO1((InvalidObjectException)value);
+ }
+ break;
+
+ case O2:
+ if (value == null) {
+ unsetO2();
+ } else {
+ setO2((AlreadyExistsException)value);
+ }
+ break;
+
+ case O3:
+ if (value == null) {
+ unsetO3();
+ } else {
+ setO3((MetaException)value);
+ }
+ break;
+
+ }
+ }
+
+ public Object getFieldValue(_Fields field) {
+ switch (field) {
+ case SUCCESS:
+ return getSuccess();
+
+ case O1:
+ return getO1();
+
+ case O2:
+ return getO2();
+
+ case O3:
+ return getO3();
+
+ }
+ throw new IllegalStateException();
+ }
+
+ /** Returns true if field corresponding to fieldID is set (has been assigned a value) and false otherwise */
+ public boolean isSet(_Fields field) {
+ if (field == null) {
+ throw new IllegalArgumentException();
+ }
+
+ switch (field) {
+ case SUCCESS:
+ return isSetSuccess();
+ case O1:
+ return isSetO1();
+ case O2:
+ return isSetO2();
+ case O3:
+ return isSetO3();
+ }
+ throw new IllegalStateException();
+ }
+
+ @Override
+ public boolean equals(Object that) {
+ if (that == null)
+ return false;
+ if (that instanceof append_partition_result)
+ return this.equals((append_partition_result)that);
+ return false;
+ }
+
+ public boolean equals(append_partition_result that) {
+ if (that == null)
+ return false;
+
+ boolean this_present_success = true && this.isSetSuccess();
+ boolean that_present_success = true && that.isSetSuccess();
+ if (this_present_success || that_present_success) {
+ if (!(this_present_success && that_present_success))
+ return false;
+ if (!this.success.equals(that.success))
+ return false;
+ }
+
+ boolean this_present_o1 = true && this.isSetO1();
+ boolean that_present_o1 = true && that.isSetO1();
+ if (this_present_o1 || that_present_o1) {
+ if (!(this_present_o1 && that_present_o1))
+ return false;
+ if (!this.o1.equals(that.o1))
+ return false;
+ }
+
+ boolean this_present_o2 = true && this.isSetO2();
+ boolean that_present_o2 = true && that.isSetO2();
+ if (this_present_o2 || that_present_o2) {
+ if (!(this_present_o2 && that_present_o2))
+ return false;
+ if (!this.o2.equals(that.o2))
+ return false;
+ }
+
+ boolean this_present_o3 = true && this.isSetO3();
+ boolean that_present_o3 = true && that.isSetO3();
+ if (this_present_o3 || that_present_o3) {
+ if (!(this_present_o3 && that_present_o3))
+ return false;
+ if (!this.o3.equals(that.o3))
+ return false;
+ }
+
+ return true;
+ }
+
+ @Override
+ public int hashCode() {
+ HashCodeBuilder builder = new HashCodeBuilder();
+
+ boolean present_success = true && (isSetSuccess());
+ builder.append(present_success);
+ if (present_success)
+ builder.append(success);
+
+ boolean present_o1 = true && (isSetO1());
+ builder.append(present_o1);
+ if (present_o1)
+ builder.append(o1);
+
+ boolean present_o2 = true && (isSetO2());
+ builder.append(present_o2);
+ if (present_o2)
+ builder.append(o2);
+
+ boolean present_o3 = true && (isSetO3());
+ builder.append(present_o3);
+ if (present_o3)
+ builder.append(o3);
+
+ return builder.toHashCode();
+ }
+
+ public int compareTo(append_partition_result other) {
+ if (!getClass().equals(other.getClass())) {
+ return getClass().getName().compareTo(other.getClass().getName());
+ }
+
+ int lastComparison = 0;
+ append_partition_result typedOther = (append_partition_result)other;
+
+ lastComparison = Boolean.valueOf(isSetSuccess()).compareTo(typedOther.isSetSuccess());
+ if (lastComparison != 0) {
+ return lastComparison;
+ }
+ if (isSetSuccess()) {
+ lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.success, typedOther.success);
+ if (lastComparison != 0) {
+ return lastComparison;
+ }
+ }
+ lastComparison = Boolean.valueOf(isSetO1()).compareTo(typedOther.isSetO1());
+ if (lastComparison != 0) {
+ return lastComparison;
+ }
+ if (isSetO1()) {
+ lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.o1, typedOther.o1);
+ if (lastComparison != 0) {
+ return lastComparison;
+ }
+ }
+ lastComparison = Boolean.valueOf(isSetO2()).compareTo(typedOther.isSetO2());
+ if (lastComparison != 0) {
+ return lastComparison;
+ }
+ if (isSetO2()) {
+ lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.o2, typedOther.o2);
+ if (lastComparison != 0) {
+ return lastComparison;
+ }
+ }
+ lastComparison = Boolean.valueOf(isSetO3()).compareTo(typedOther.isSetO3());
+ if (lastComparison != 0) {
+ return lastComparison;
+ }
+ if (isSetO3()) {
+ lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.o3, typedOther.o3);
+ if (lastComparison != 0) {
+ return lastComparison;
+ }
+ }
+ return 0;
+ }
+
+ public _Fields fieldForId(int fieldId) {
+ return _Fields.findByThriftId(fieldId);
+ }
+
+ public void read(org.apache.thrift.protocol.TProtocol iprot) throws org.apache.thrift.TException {
+ schemes.get(iprot.getScheme()).getScheme().read(iprot, this);
+ }
+
+ public void write(org.apache.thrift.protocol.TProtocol oprot) throws org.apache.thrift.TException {
+ schemes.get(oprot.getScheme()).getScheme().write(oprot, this);
+ }
+
+ @Override
+ public String toString() {
+ StringBuilder sb = new StringBuilder("append_partition_result(");
+ boolean first = true;
+
+ sb.append("success:");
+ if (this.success == null) {
+ sb.append("null");
+ } else {
+ sb.append(this.success);
+ }
+ first = false;
+ if (!first) sb.append(", ");
+ sb.append("o1:");
+ if (this.o1 == null) {
+ sb.append("null");
+ } else {
+ sb.append(this.o1);
+ }
+ first = false;
+ if (!first) sb.append(", ");
+ sb.append("o2:");
+ if (this.o2 == null) {
+ sb.append("null");
+ } else {
+ sb.append(this.o2);
+ }
+ first = false;
+ if (!first) sb.append(", ");
+ sb.append("o3:");
+ if (this.o3 == null) {
+ sb.append("null");
+ } else {
+ sb.append(this.o3);
+ }
+ first = false;
+ sb.append(")");
+ return sb.toString();
+ }
+
+ public void validate() throws org.apache.thrift.TException {
+ // check for required fields
+ // check for sub-struct validity
+ if (success != null) {
+ success.validate();
+ }
+ }
+
+ private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException {
+ try {
+ write(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(out)));
+ } catch (org.apache.thrift.TException te) {
+ throw new java.io.IOException(te);
+ }
+ }
+
+ private void readObject(java.io.ObjectInputStream in) throws java.io.IOException, ClassNotFoundException {
+ try {
+ read(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(in)));
+ } catch (org.apache.thrift.TException te) {
+ throw new java.io.IOException(te);
+ }
+ }
+
+ private static class append_partition_resultStandardSchemeFactory implements SchemeFactory {
+ public append_partition_resultStandardScheme getScheme() {
+ return new append_partition_resultStandardScheme();
+ }
+ }
+
+ private static class append_partition_resultStandardScheme extends StandardScheme<append_partition_result> {
+
+ public void read(org.apache.thrift.protocol.TProtocol iprot, append_partition_result struct) throws org.apache.thrift.TException {
+ org.apache.thrift.protocol.TField schemeField;
+ iprot.readStructBegin();
+ while (true)
+ {
+ schemeField = iprot.readFieldBegin();
+ if (schemeField.type == org.apache.thrift.protocol.TType.STOP) {
+ break;
+ }
+ switch (schemeField.id) {
+ case 0: // SUCCESS
+ if (schemeField.type == org.apache.thrift.protocol.TType.STRUCT) {
+ struct.success = new Partition();
+ struct.success.read(iprot);
+ struct.setSuccessIsSet(true);
+ } else {
+ org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type);
+ }
+ break;
+ case 1: // O1
+ if (schemeField.type == org.apache.thrift.protocol.TType.STRUCT) {
+ struct.o1 = new InvalidObjectException();
+ struct.o1.read(iprot);
+ struct.setO1IsSet(true);
+ } else {
+ org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type);
+ }
+ break;
+ case 2: // O2
+ if (schemeField.type == org.apache.thrift.protocol.TType.STRUCT) {
+ struct.o2 = new AlreadyExistsException();
+ struct.o2.read(iprot);
+ struct.setO2IsSet(true);
+ } else {
+ org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type);
+ }
+ break;
+ case 3: // O3
+ if (schemeField.type == org.apache.thrift.protocol.TType.STRUCT) {
+ struct.o3 = new MetaException();
+ struct.o3.read(iprot);
+ struct.setO3IsSet(true);
+ } else {
+ org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type);
+ }
+ break;
+ default:
+ org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type);
+ }
+ iprot.readFieldEnd();
+ }
+ iprot.readStructEnd();
+ struct.validate();
+ }
+
+ public void write(org.apache.thrift.protocol.TProtocol oprot, append_partition_result struct) throws org.apache.thrift.TException {
+ struct.validate();
+
+ oprot.writeStructBegin(STRUCT_DESC);
+ if (struct.success != null) {
+ oprot.writeFieldBegin(SUCCESS_FIELD_DESC);
+ struct.success.write(oprot);
+ oprot.writeFieldEnd();
+ }
+ if (struct.o1 != null) {
+ oprot.writeFieldBegin(O1_FIELD_DESC);
+ struct.o1.write(oprot);
+ oprot.writeFieldEnd();
+ }
+ if (struct.o2 != null) {
+ oprot.writeFieldBegin(O2_FIELD_DESC);
+ struct.o2.write(oprot);
+ oprot.writeFieldEnd();
+ }
+ if (struct.o3 != null) {
+ oprot.writeFieldBegin(O3_FIELD_DESC);
+ struct.o3.write(oprot);
+ oprot.writeFieldEnd();
+ }
+ oprot.writeFieldStop();
+ oprot.writeStructEnd();
+ }
+
+ }
+
+ private static class append_partition_resultTupleSchemeFactory implements SchemeFactory {
+ public append_partition_resultTupleScheme getScheme() {
+ return new append_partition_resultTupleScheme();
+ }
+ }
+
+ private static class append_partition_resultTupleScheme extends TupleScheme<append_partition_result> {
+
+ @Override
+ public void write(org.apache.thrift.protocol.TProtocol prot, append_partition_result struct) throws org.apache.thrift.TException {
+ TTupleProtocol oprot = (TTupleProtocol) prot;
+ BitSet optionals = new BitSet();
+ if (struct.isSetSuccess()) {
+ optionals.set(0);
+ }
+ if (struct.isSetO1()) {
+ optionals.set(1);
+ }
+ if (struct.isSetO2()) {
+ optionals.set(2);
+ }
+ if (struct.isSetO3()) {
+ optionals.set(3);
+ }
+ oprot.writeBitSet(optionals, 4);
+ if (struct.isSetSuccess()) {
+ struct.success.write(oprot);
+ }
+ if (struct.isSetO1()) {
+ struct.o1.write(oprot);
+ }
+ if (struct.isSetO2()) {
+ struct.o2.write(oprot);
+ }
+ if (struct.isSetO3()) {
+ struct.o3.write(oprot);
+ }
+ }
+
+ @Override
+ public void read(org.apache.thrift.protocol.TProtocol prot, append_partition_result struct) throws org.apache.thrift.TException {
+ TTupleProtocol iprot = (TTupleProtocol) prot;
+ BitSet incoming = iprot.readBitSet(4);
+ if (incoming.get(0)) {
+ struct.success = new Partition();
+ struct.success.read(iprot);
+ struct.setSuccessIsSet(true);
+ }
+ if (incoming.get(1)) {
+ struct.o1 = new InvalidObjectException();
+ struct.o1.read(iprot);
+ struct.setO1IsSet(true);
+ }
+ if (incoming.get(2)) {
+ struct.o2 = new AlreadyExistsException();
+ struct.o2.read(iprot);
+ struct.setO2IsSet(true);
+ }
+ if (incoming.get(3)) {
+ struct.o3 = new MetaException();
+ struct.o3.read(iprot);
+ struct.setO3IsSet(true);
+ }
+ }
+ }
+
+ }
+
+ public static class add_partitions_req_args implements org.apache.thrift.TBase<add_partitions_req_args, add_partitions_req_args._Fields>, java.io.Serializable, Cloneable {
+ private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("add_partitions_req_args");
+
+ private static final org.apache.thrift.protocol.TField REQUEST_FIELD_DESC = new org.apache.thrift.protocol.TField("request", org.apache.thrift.protocol.TType.STRUCT, (short)1);
+
+ private static final Map<Class<? extends IScheme>, SchemeFactory> schemes = new HashMap<Class<? extends IScheme>, SchemeFactory>();
+ static {
+ schemes.put(StandardScheme.class, new add_partitions_req_argsStandardSchemeFactory());
+ schemes.put(TupleScheme.class, new add_partitions_req_argsTupleSchemeFactory());
+ }
+
+ private AddPartitionsRequest request; // required
+
+ /** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */
+ public enum _Fields implements org.apache.thrift.TFieldIdEnum {
+ REQUEST((short)1, "request");
+
+ private static final Map<String, _Fields> byName = new HashMap<String, _Fields>();
+
+ static {
+ for (_Fields field : EnumSet.allOf(_Fields.class)) {
+ byName.put(field.getFieldName(), field);
+ }
+ }
+
+ /**
+ * Find the _Fields constant that matches fieldId, or null if its not found.
+ */
+ public static _Fields findByThriftId(int fieldId) {
+ switch(fieldId) {
+ case 1: // REQUEST
+ return REQUEST;
+ default:
+ return null;
+ }
+ }
+
+ /**
+ * Find the _Fields constant that matches fieldId, throwing an exception
+ * if it is not found.
+ */
+ public static _Fields findByThriftIdOrThrow(int fieldId) {
+ _Fields fields = findByThriftId(fieldId);
+ if (fields == null) throw new IllegalArgumentException("Field " + fieldId + " doesn't exist!");
+ return fields;
+ }
+
+ /**
+ * Find the _Fields constant that matches name, or null if its not found.
+ */
+ public static _Fields findByName(String name) {
+ return byName.get(name);
+ }
+
+ private final short _thriftId;
+ private final String _fieldName;
+
+ _Fields(short thriftId, String fieldName) {
+ _thriftId = thriftId;
+ _fieldName = fieldName;
+ }
+
+ public short getThriftFieldId() {
+ return _thriftId;
+ }
+
+ public String getFieldName() {
+ return _fieldName;
+ }
+ }
+
+ // isset id assignments
+ public static final Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap;
+ static {
+ Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class);
+ tmpMap.put(_Fields.REQUEST, new org.apache.thrift.meta_data.FieldMetaData("request", org.apache.thrift.TFieldRequirementType.DEFAULT,
+ new org.apache.thrift.meta_data.StructMetaData(org.apache.thrift.protocol.TType.STRUCT, AddPartitionsRequest.class)));
+ metaDataMap = Collections.unmodifiableMap(tmpMap);
+ org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(add_partitions_req_args.class, metaDataMap);
+ }
+
+ public add_partitions_req_args() {
+ }
+
+ public add_partitions_req_args(
+ AddPartitionsRequest request)
+ {
+ this();
+ this.request = request;
+ }
+
+ /**
+ * Performs a deep copy on <i>other</i>.
+ */
+ public add_partitions_req_args(add_partitions_req_args other) {
+ if (other.isSetRequest()) {
+ this.request = new AddPartitionsRequest(other.request);
+ }
+ }
+
+ public add_partitions_req_args deepCopy() {
+ return new add_partitions_req_args(this);
+ }
+
+ @Override
+ public void clear() {
+ this.request = null;
+ }
+
+ public AddPartitionsRequest getRequest() {
+ return this.request;
+ }
+
+ public void setRequest(AddPartitionsRequest request) {
+ this.request = request;
+ }
+
+ public void unsetRequest() {
+ this.request = null;
+ }
+
+ /** Returns true if field request is set (has been assigned a value) and false otherwise */
+ public boolean isSetRequest() {
+ return this.request != null;
+ }
+
+ public void setRequestIsSet(boolean value) {
+ if (!value) {
+ this.request = null;
+ }
+ }
+
+ public void setFieldValue(_Fields field, Object value) {
+ switch (field) {
+ case REQUEST:
+ if (value == null) {
+ unsetRequest();
+ } else {
+ setRequest((AddPartitionsRequest)value);
+ }
+ break;
+
+ }
+ }
+
+ public Object getFieldValue(_Fields field) {
+ switch (field) {
+ case REQUEST:
+ return getRequest();
+
+ }
+ throw new IllegalStateException();
+ }
+
+ /** Returns true if field corresponding to fieldID is set (has been assigned a value) and false otherwise */
+ public boolean isSet(_Fields field) {
+ if (field == null) {
+ throw new IllegalArgumentException();
+ }
+
+ switch (field) {
+ case REQUEST:
+ return isSetRequest();
+ }
+ throw new IllegalStateException();
+ }
+
+ @Override
+ public boolean equals(Object that) {
+ if (that == null)
+ return false;
+ if (that instanceof add_partitions_req_args)
+ return this.equals((add_partitions_req_args)that);
+ return false;
+ }
+
+ public boolean equals(add_partitions_req_args that) {
+ if (that == null)
+ return false;
+
+ boolean this_present_request = true && this.isSetRequest();
+ boolean that_present_request = true && that.isSetRequest();
+ if (this_present_request || that_present_request) {
+ if (!(this_present_request && that_present_request))
+ return false;
+ if (!this.request.equals(that.request))
+ return false;
+ }
+
+ return true;
+ }
+
+ @Override
+ public int hashCode() {
+ HashCodeBuilder builder = new HashCodeBuilder();
+
+ boolean present_request = true && (isSetRequest());
+ builder.append(present_request);
+ if (present_request)
+ builder.append(request);
+
+ return builder.toHashCode();
+ }
+
+ public int compareTo(add_partitions_req_args other) {
+ if (!getClass().equals(other.getClass())) {
+ return getClass().getName().compareTo(other.getClass().getName());
+ }
+
+ int lastComparison = 0;
+ add_partitions_req_args typedOther = (add_partitions_req_args)other;
+
+ lastComparison = Boolean.valueOf(isSetRequest()).compareTo(typedOther.isSetRequest());
+ if (lastComparison != 0) {
+ return lastComparison;
+ }
+ if (isSetRequest()) {
+ lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.request, typedOther.request);
+ if (lastComparison != 0) {
+ return lastComparison;
+ }
+ }
+ return 0;
+ }
+
+ public _Fields fieldForId(int fieldId) {
+ return _Fields.findByThriftId(fieldId);
+ }
+
+ public void read(org.apache.thrift.protocol.TProtocol iprot) throws org.apache.thrift.TException {
+ schemes.get(iprot.getScheme()).getScheme().read(iprot, this);
+ }
+
+ public void write(org.apache.thrift.protocol.TProtocol oprot) throws org.apache.thrift.TException {
+ schemes.get(oprot.getScheme()).getScheme().write(oprot, this);
+ }
+
+ @Override
+ public String toString() {
+ StringBuilder sb = new StringBuilder("add_partitions_req_args(");
+ boolean first = true;
+
+ sb.append("request:");
+ if (this.request == null) {
+ sb.append("null");
+ } else {
+ sb.append(this.request);
+ }
+ first = false;
+ sb.append(")");
+ return sb.toString();
+ }
+
+ public void validate() throws org.apache.thrift.TException {
+ // check for required fields
+ // check for sub-struct validity
+ if (request != null) {
+ request.validate();
+ }
+ }
+
+ private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException {
+ try {
+ write(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(out)));
+ } catch (org.apache.thrift.TException te) {
+ throw new java.io.IOException(te);
+ }
+ }
+
+ private void readObject(java.io.ObjectInputStream in) throws java.io.IOException, ClassNotFoundException {
+ try {
+ read(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(in)));
+ } catch (org.apache.thrift.TException te) {
+ throw new java.io.IOException(te);
+ }
+ }
+
+ private static class add_partitions_req_argsStandardSchemeFactory implements SchemeFactory {
+ public add_partitions_req_argsStandardScheme getScheme() {
+ return new add_partitions_req_argsStandardScheme();
+ }
+ }
+
+ private static class add_partitions_req_argsStandardScheme extends StandardScheme<add_partitions_req_args> {
+
+ public void read(org.apache.thrift.protocol.TProtocol iprot, add_partitions_req_args struct) throws org.apache.thrift.TException {
+ org.apache.thrift.protocol.TField schemeField;
+ iprot.readStructBegin();
+ while (true)
+ {
+ schemeField = iprot.readFieldBegin();
+ if (schemeField.type == org.apache.thrift.protocol.TType.STOP) {
+ break;
+ }
+ switch (schemeField.id) {
+ case 1: // REQUEST
+ if (schemeField.type == org.apache.thrift.protocol.TType.STRUCT) {
+ struct.request = new AddPartitionsRequest();
+ struct.request.read(iprot);
+ struct.setRequestIsSet(true);
+ } else {
+ org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type);
+ }
+ break;
+ default:
+ org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type);
+ }
+ iprot.readFieldEnd();
+ }
+ iprot.readStructEnd();
+ struct.validate();
+ }
+
+ public void write(org.apache.thrift.protocol.TProtocol oprot, add_partitions_req_args struct) throws org.apache.thrift.TException {
+ struct.validate();
+
+ oprot.writeStructBegin(STRUCT_DESC);
+ if (struct.request != null) {
+ oprot.writeFieldBegin(REQUEST_FIELD_DESC);
+ struct.request.write(oprot);
oprot.writeFieldEnd();
}
oprot.writeFieldStop();
@@ -35871,77 +37046,43 @@ public class ThriftHiveMetastore {
}
- private static class append_partition_argsTupleSchemeFactory implements SchemeFactory {
- public append_partition_argsTupleScheme getScheme() {
- return new append_partition_argsTupleScheme();
+ private static class add_partitions_req_argsTupleSchemeFactory implements SchemeFactory {
+ public add_partitions_req_argsTupleScheme getScheme() {
+ return new add_partitions_req_argsTupleScheme();
}
}
- private static class append_partition_argsTupleScheme extends TupleScheme<append_partition_args> {
+ private static class add_partitions_req_argsTupleScheme extends TupleScheme<add_partitions_req_args> {
@Override
- public void write(org.apache.thrift.protocol.TProtocol prot, append_partition_args struct) throws org.apache.thrift.TException {
+ public void write(org.apache.thrift.protocol.TProtocol prot, add_partitions_req_args struct) throws org.apache.thrift.TException {
TTupleProtocol oprot = (TTupleProtocol) prot;
BitSet optionals = new BitSet();
- if (struct.isSetDb_name()) {
+ if (struct.isSetRequest()) {
optionals.set(0);
}
- if (struct.isSetTbl_name()) {
- optionals.set(1);
- }
- if (struct.isSetPart_vals()) {
- optionals.set(2);
- }
- oprot.writeBitSet(optionals, 3);
- if (struct.isSetDb_name()) {
- oprot.writeString(struct.db_name);
- }
- if (struct.isSetTbl_name()) {
- oprot.writeString(struct.tbl_name);
- }
- if (struct.isSetPart_vals()) {
- {
- oprot.writeI32(struct.part_vals.size());
- for (String _iter358 : struct.part_vals)
- {
- oprot.writeString(_iter358);
- }
- }
+ oprot.writeBitSet(optionals, 1);
+ if (struct.isSetRequest()) {
+ struct.request.write(oprot);
}
}
@Override
- public void read(org.apache.thrift.protocol.TProtocol prot, append_partition_args struct) throws org.apache.thrift.TException {
+ public void read(org.apache.thrift.protocol.TProtocol prot, add_partitions_req_args struct) throws org.apache.thrift.TException {
TTupleProtocol iprot = (TTupleProtocol) prot;
- BitSet incoming = iprot.readBitSet(3);
+ BitSet incoming = iprot.readBitSet(1);
if (incoming.get(0)) {
- struct.db_name = iprot.readString();
- struct.setDb_nameIsSet(true);
- }
- if (incoming.get(1)) {
- struct.tbl_name = iprot.readString();
- struct.setTbl_nameIsSet(true);
- }
- if (incoming.get(2)) {
- {
- org.apache.thrift.protocol.TList _list359 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
- struct.part_vals = new ArrayList<String>(_list359.size);
- for (int _i360 = 0; _i360 < _list359.size; ++_i360)
- {
- String _elem361; // required
- _elem361 = iprot.readString();
- struct.part_vals.add(_elem361);
- }
- }
- struct.setPart_valsIsSet(true);
+ struct.request = new AddPartitionsRequest();
+ struct.request.read(iprot);
+ struct.setRequestIsSet(true);
}
}
}
}
- public static class append_partition_result implements org.apache.thrift.TBase<append_partition_result, append_partition_result._Fields>, java.io.Serializable, Cloneable {
- private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("append_partition_result");
+ public static class add_partitions_req_result implements org.apache.thrift.TBase<add_partitions_req_result, add_partitions_req_result._Fields>, java.io.Serializable, Cloneable {
+ private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("add_partitions_req_result");
private static final org.apache.thrift.protocol.TField SUCCESS_FIELD_DESC = new org.apache.thrift.protocol.TField("success", org.apache.thrift.protocol.TType.STRUCT, (short)0);
private static final org.apache.thrift.protocol.TField O1_FIELD_DESC = new org.apache.thrift.protocol.TField("o1", org.apache.thrift.protocol.TType.STRUCT, (short)1);
@@ -35950,11 +37091,11 @@ public class ThriftHiveMetastore {
private static final Map<Class<? extends IScheme>, SchemeFactory> schemes = new HashMap<Class<? extends IScheme>, SchemeFactory>();
static {
- schemes.put(StandardScheme.class, new append_partition_resultStandardSchemeFactory());
- schemes.put(TupleScheme.class, new append_partition_resultTupleSchemeFactory());
+ schemes.put(StandardScheme.class, new add_partitions_req_resultStandardSchemeFactory());
+ schemes.put(TupleScheme.class, new add_partitions_req_resultTupleSchemeFactory());
}
- private Partition success; // required
+ private AddPartitionsResult success; // required
private InvalidObjectException o1; // required
private AlreadyExistsException o2; // required
private MetaException o3; // required
@@ -36031,7 +37172,7 @@ public class ThriftHiveMetastore {
static {
Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class);
tmpMap.put(_Fields.SUCCESS, new org.apache.thrift.meta_data.FieldMetaData("success", org.apache.thrift.TFieldRequirementType.DEFAULT,
- new org.apache.thrift.meta_data.StructMetaData(org.apache.thrift.protocol.TType.STRUCT, Partition.class)));
+ new org.apache.thrift.meta_data.StructMetaData(org.apache.thrift.protocol.TType.STRUCT, AddPartitionsResult.class)));
tmpMap.put(_Fields.O1, new org.apache.thrift.meta_data.FieldMetaData("o1", org.apache.thrift.TFieldRequirementType.DEFAULT,
new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRUCT)));
tmpMap.put(_Fields.O2, new org.apache.thrift.meta_data.FieldMetaData("o2", org.apache.thrift.TFieldRequirementType.DEFAULT,
@@ -36039,14 +37180,14 @@ public class ThriftHiveMetastore {
tmpMap.put(_Fields.O3, new org.apache.thrift.meta_data.FieldMetaData("o3", org.apache.thrift.TFieldRequirementType.DEFAULT,
new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRUCT)));
metaDataMap = Collections.unmodifiableMap(tmpMap);
- org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(append_partition_result.class, metaDataMap);
[... 2569 lines stripped ...]