You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by gu...@apache.org on 2014/09/05 21:15:48 UTC
svn commit: r1622763 [8/17] - in /hive/branches/tez: ./
ant/src/org/apache/hadoop/hive/ant/
hcatalog/streaming/src/java/org/apache/hive/hcatalog/streaming/
hcatalog/streaming/src/test/org/apache/hive/hcatalog/streaming/
itests/hive-unit/ itests/hive-un...
Modified: hive/branches/tez/metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ThriftHiveMetastore.java
URL: http://svn.apache.org/viewvc/hive/branches/tez/metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ThriftHiveMetastore.java?rev=1622763&r1=1622762&r2=1622763&view=diff
==============================================================================
--- hive/branches/tez/metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ThriftHiveMetastore.java (original)
+++ hive/branches/tez/metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ThriftHiveMetastore.java Fri Sep 5 19:15:44 2014
@@ -94,6 +94,8 @@ public class ThriftHiveMetastore {
public int add_partitions(List<Partition> new_parts) throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException;
+ public int add_partitions_pspec(List<PartitionSpec> new_parts) throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException;
+
public Partition append_partition(String db_name, String tbl_name, List<String> part_vals) throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException;
public AddPartitionsResult add_partitions_req(AddPartitionsRequest request) throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException;
@@ -126,6 +128,8 @@ public class ThriftHiveMetastore {
public List<Partition> get_partitions_with_auth(String db_name, String tbl_name, short max_parts, String user_name, List<String> group_names) throws NoSuchObjectException, MetaException, org.apache.thrift.TException;
+ public List<PartitionSpec> get_partitions_pspec(String db_name, String tbl_name, int max_parts) throws NoSuchObjectException, MetaException, org.apache.thrift.TException;
+
public List<String> get_partition_names(String db_name, String tbl_name, short max_parts) throws MetaException, org.apache.thrift.TException;
public List<Partition> get_partitions_ps(String db_name, String tbl_name, List<String> part_vals, short max_parts) throws MetaException, NoSuchObjectException, org.apache.thrift.TException;
@@ -136,6 +140,8 @@ public class ThriftHiveMetastore {
public List<Partition> get_partitions_by_filter(String db_name, String tbl_name, String filter, short max_parts) throws MetaException, NoSuchObjectException, org.apache.thrift.TException;
+ public List<PartitionSpec> get_part_specs_by_filter(String db_name, String tbl_name, String filter, int max_parts) throws MetaException, NoSuchObjectException, org.apache.thrift.TException;
+
public PartitionsByExprResult get_partitions_by_expr(PartitionsByExprRequest req) throws MetaException, NoSuchObjectException, org.apache.thrift.TException;
public List<Partition> get_partitions_by_names(String db_name, String tbl_name, List<String> names) throws MetaException, NoSuchObjectException, org.apache.thrift.TException;
@@ -324,6 +330,8 @@ public class ThriftHiveMetastore {
public void add_partitions(List<Partition> new_parts, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.add_partitions_call> resultHandler) throws org.apache.thrift.TException;
+ public void add_partitions_pspec(List<PartitionSpec> new_parts, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.add_partitions_pspec_call> resultHandler) throws org.apache.thrift.TException;
+
public void append_partition(String db_name, String tbl_name, List<String> part_vals, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.append_partition_call> resultHandler) throws org.apache.thrift.TException;
public void add_partitions_req(AddPartitionsRequest request, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.add_partitions_req_call> resultHandler) throws org.apache.thrift.TException;
@@ -356,6 +364,8 @@ public class ThriftHiveMetastore {
public void get_partitions_with_auth(String db_name, String tbl_name, short max_parts, String user_name, List<String> group_names, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.get_partitions_with_auth_call> resultHandler) throws org.apache.thrift.TException;
+ public void get_partitions_pspec(String db_name, String tbl_name, int max_parts, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.get_partitions_pspec_call> resultHandler) throws org.apache.thrift.TException;
+
public void get_partition_names(String db_name, String tbl_name, short max_parts, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.get_partition_names_call> resultHandler) throws org.apache.thrift.TException;
public void get_partitions_ps(String db_name, String tbl_name, List<String> part_vals, short max_parts, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.get_partitions_ps_call> resultHandler) throws org.apache.thrift.TException;
@@ -366,6 +376,8 @@ public class ThriftHiveMetastore {
public void get_partitions_by_filter(String db_name, String tbl_name, String filter, short max_parts, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.get_partitions_by_filter_call> resultHandler) throws org.apache.thrift.TException;
+ public void get_part_specs_by_filter(String db_name, String tbl_name, String filter, int max_parts, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.get_part_specs_by_filter_call> resultHandler) throws org.apache.thrift.TException;
+
public void get_partitions_by_expr(PartitionsByExprRequest req, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.get_partitions_by_expr_call> resultHandler) throws org.apache.thrift.TException;
public void get_partitions_by_names(String db_name, String tbl_name, List<String> names, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.get_partitions_by_names_call> resultHandler) throws org.apache.thrift.TException;
@@ -1341,6 +1353,38 @@ public class ThriftHiveMetastore {
throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "add_partitions failed: unknown result");
}
+ public int add_partitions_pspec(List<PartitionSpec> new_parts) throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException
+ {
+ send_add_partitions_pspec(new_parts);
+ return recv_add_partitions_pspec();
+ }
+
+ public void send_add_partitions_pspec(List<PartitionSpec> new_parts) throws org.apache.thrift.TException
+ {
+ add_partitions_pspec_args args = new add_partitions_pspec_args();
+ args.setNew_parts(new_parts);
+ sendBase("add_partitions_pspec", args);
+ }
+
+ public int recv_add_partitions_pspec() throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException
+ {
+ add_partitions_pspec_result result = new add_partitions_pspec_result();
+ receiveBase(result, "add_partitions_pspec");
+ if (result.isSetSuccess()) {
+ return result.success;
+ }
+ if (result.o1 != null) {
+ throw result.o1;
+ }
+ if (result.o2 != null) {
+ throw result.o2;
+ }
+ if (result.o3 != null) {
+ throw result.o3;
+ }
+ throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "add_partitions_pspec failed: unknown result");
+ }
+
public Partition append_partition(String db_name, String tbl_name, List<String> part_vals) throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException
{
send_append_partition(db_name, tbl_name, part_vals);
@@ -1868,6 +1912,37 @@ public class ThriftHiveMetastore {
throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "get_partitions_with_auth failed: unknown result");
}
+ public List<PartitionSpec> get_partitions_pspec(String db_name, String tbl_name, int max_parts) throws NoSuchObjectException, MetaException, org.apache.thrift.TException
+ {
+ send_get_partitions_pspec(db_name, tbl_name, max_parts);
+ return recv_get_partitions_pspec();
+ }
+
+ public void send_get_partitions_pspec(String db_name, String tbl_name, int max_parts) throws org.apache.thrift.TException
+ {
+ get_partitions_pspec_args args = new get_partitions_pspec_args();
+ args.setDb_name(db_name);
+ args.setTbl_name(tbl_name);
+ args.setMax_parts(max_parts);
+ sendBase("get_partitions_pspec", args);
+ }
+
+ public List<PartitionSpec> recv_get_partitions_pspec() throws NoSuchObjectException, MetaException, org.apache.thrift.TException
+ {
+ get_partitions_pspec_result result = new get_partitions_pspec_result();
+ receiveBase(result, "get_partitions_pspec");
+ if (result.isSetSuccess()) {
+ return result.success;
+ }
+ if (result.o1 != null) {
+ throw result.o1;
+ }
+ if (result.o2 != null) {
+ throw result.o2;
+ }
+ throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "get_partitions_pspec failed: unknown result");
+ }
+
public List<String> get_partition_names(String db_name, String tbl_name, short max_parts) throws MetaException, org.apache.thrift.TException
{
send_get_partition_names(db_name, tbl_name, max_parts);
@@ -2026,6 +2101,38 @@ public class ThriftHiveMetastore {
throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "get_partitions_by_filter failed: unknown result");
}
+ public List<PartitionSpec> get_part_specs_by_filter(String db_name, String tbl_name, String filter, int max_parts) throws MetaException, NoSuchObjectException, org.apache.thrift.TException
+ {
+ send_get_part_specs_by_filter(db_name, tbl_name, filter, max_parts);
+ return recv_get_part_specs_by_filter();
+ }
+
+ public void send_get_part_specs_by_filter(String db_name, String tbl_name, String filter, int max_parts) throws org.apache.thrift.TException
+ {
+ get_part_specs_by_filter_args args = new get_part_specs_by_filter_args();
+ args.setDb_name(db_name);
+ args.setTbl_name(tbl_name);
+ args.setFilter(filter);
+ args.setMax_parts(max_parts);
+ sendBase("get_part_specs_by_filter", args);
+ }
+
+ public List<PartitionSpec> recv_get_part_specs_by_filter() throws MetaException, NoSuchObjectException, org.apache.thrift.TException
+ {
+ get_part_specs_by_filter_result result = new get_part_specs_by_filter_result();
+ receiveBase(result, "get_part_specs_by_filter");
+ if (result.isSetSuccess()) {
+ return result.success;
+ }
+ if (result.o1 != null) {
+ throw result.o1;
+ }
+ if (result.o2 != null) {
+ throw result.o2;
+ }
+ throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "get_part_specs_by_filter failed: unknown result");
+ }
+
public PartitionsByExprResult get_partitions_by_expr(PartitionsByExprRequest req) throws MetaException, NoSuchObjectException, org.apache.thrift.TException
{
send_get_partitions_by_expr(req);
@@ -4840,6 +4947,38 @@ public class ThriftHiveMetastore {
}
}
+ public void add_partitions_pspec(List<PartitionSpec> new_parts, org.apache.thrift.async.AsyncMethodCallback<add_partitions_pspec_call> resultHandler) throws org.apache.thrift.TException {
+ checkReady();
+ add_partitions_pspec_call method_call = new add_partitions_pspec_call(new_parts, resultHandler, this, ___protocolFactory, ___transport);
+ this.___currentMethod = method_call;
+ ___manager.call(method_call);
+ }
+
+ public static class add_partitions_pspec_call extends org.apache.thrift.async.TAsyncMethodCall {
+ private List<PartitionSpec> new_parts;
+ public add_partitions_pspec_call(List<PartitionSpec> new_parts, org.apache.thrift.async.AsyncMethodCallback<add_partitions_pspec_call> resultHandler, org.apache.thrift.async.TAsyncClient client, org.apache.thrift.protocol.TProtocolFactory protocolFactory, org.apache.thrift.transport.TNonblockingTransport transport) throws org.apache.thrift.TException {
+ super(client, protocolFactory, transport, resultHandler, false);
+ this.new_parts = new_parts;
+ }
+
+ public void write_args(org.apache.thrift.protocol.TProtocol prot) throws org.apache.thrift.TException {
+ prot.writeMessageBegin(new org.apache.thrift.protocol.TMessage("add_partitions_pspec", org.apache.thrift.protocol.TMessageType.CALL, 0));
+ add_partitions_pspec_args args = new add_partitions_pspec_args();
+ args.setNew_parts(new_parts);
+ args.write(prot);
+ prot.writeMessageEnd();
+ }
+
+ public int getResult() throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException {
+ if (getState() != org.apache.thrift.async.TAsyncMethodCall.State.RESPONSE_READ) {
+ throw new IllegalStateException("Method call not finished!");
+ }
+ org.apache.thrift.transport.TMemoryInputTransport memoryTransport = new org.apache.thrift.transport.TMemoryInputTransport(getFrameBuffer().array());
+ org.apache.thrift.protocol.TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport);
+ return (new Client(prot)).recv_add_partitions_pspec();
+ }
+ }
+
public void append_partition(String db_name, String tbl_name, List<String> part_vals, org.apache.thrift.async.AsyncMethodCallback<append_partition_call> resultHandler) throws org.apache.thrift.TException {
checkReady();
append_partition_call method_call = new append_partition_call(db_name, tbl_name, part_vals, resultHandler, this, ___protocolFactory, ___transport);
@@ -5478,6 +5617,44 @@ public class ThriftHiveMetastore {
}
}
+ public void get_partitions_pspec(String db_name, String tbl_name, int max_parts, org.apache.thrift.async.AsyncMethodCallback<get_partitions_pspec_call> resultHandler) throws org.apache.thrift.TException {
+ checkReady();
+ get_partitions_pspec_call method_call = new get_partitions_pspec_call(db_name, tbl_name, max_parts, resultHandler, this, ___protocolFactory, ___transport);
+ this.___currentMethod = method_call;
+ ___manager.call(method_call);
+ }
+
+ public static class get_partitions_pspec_call extends org.apache.thrift.async.TAsyncMethodCall {
+ private String db_name;
+ private String tbl_name;
+ private int max_parts;
+ public get_partitions_pspec_call(String db_name, String tbl_name, int max_parts, org.apache.thrift.async.AsyncMethodCallback<get_partitions_pspec_call> resultHandler, org.apache.thrift.async.TAsyncClient client, org.apache.thrift.protocol.TProtocolFactory protocolFactory, org.apache.thrift.transport.TNonblockingTransport transport) throws org.apache.thrift.TException {
+ super(client, protocolFactory, transport, resultHandler, false);
+ this.db_name = db_name;
+ this.tbl_name = tbl_name;
+ this.max_parts = max_parts;
+ }
+
+ public void write_args(org.apache.thrift.protocol.TProtocol prot) throws org.apache.thrift.TException {
+ prot.writeMessageBegin(new org.apache.thrift.protocol.TMessage("get_partitions_pspec", org.apache.thrift.protocol.TMessageType.CALL, 0));
+ get_partitions_pspec_args args = new get_partitions_pspec_args();
+ args.setDb_name(db_name);
+ args.setTbl_name(tbl_name);
+ args.setMax_parts(max_parts);
+ args.write(prot);
+ prot.writeMessageEnd();
+ }
+
+ public List<PartitionSpec> getResult() throws NoSuchObjectException, MetaException, org.apache.thrift.TException {
+ if (getState() != org.apache.thrift.async.TAsyncMethodCall.State.RESPONSE_READ) {
+ throw new IllegalStateException("Method call not finished!");
+ }
+ org.apache.thrift.transport.TMemoryInputTransport memoryTransport = new org.apache.thrift.transport.TMemoryInputTransport(getFrameBuffer().array());
+ org.apache.thrift.protocol.TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport);
+ return (new Client(prot)).recv_get_partitions_pspec();
+ }
+ }
+
public void get_partition_names(String db_name, String tbl_name, short max_parts, org.apache.thrift.async.AsyncMethodCallback<get_partition_names_call> resultHandler) throws org.apache.thrift.TException {
checkReady();
get_partition_names_call method_call = new get_partition_names_call(db_name, tbl_name, max_parts, resultHandler, this, ___protocolFactory, ___transport);
@@ -5686,6 +5863,47 @@ public class ThriftHiveMetastore {
}
}
+ public void get_part_specs_by_filter(String db_name, String tbl_name, String filter, int max_parts, org.apache.thrift.async.AsyncMethodCallback<get_part_specs_by_filter_call> resultHandler) throws org.apache.thrift.TException {
+ checkReady();
+ get_part_specs_by_filter_call method_call = new get_part_specs_by_filter_call(db_name, tbl_name, filter, max_parts, resultHandler, this, ___protocolFactory, ___transport);
+ this.___currentMethod = method_call;
+ ___manager.call(method_call);
+ }
+
+ public static class get_part_specs_by_filter_call extends org.apache.thrift.async.TAsyncMethodCall {
+ private String db_name;
+ private String tbl_name;
+ private String filter;
+ private int max_parts;
+ public get_part_specs_by_filter_call(String db_name, String tbl_name, String filter, int max_parts, org.apache.thrift.async.AsyncMethodCallback<get_part_specs_by_filter_call> resultHandler, org.apache.thrift.async.TAsyncClient client, org.apache.thrift.protocol.TProtocolFactory protocolFactory, org.apache.thrift.transport.TNonblockingTransport transport) throws org.apache.thrift.TException {
+ super(client, protocolFactory, transport, resultHandler, false);
+ this.db_name = db_name;
+ this.tbl_name = tbl_name;
+ this.filter = filter;
+ this.max_parts = max_parts;
+ }
+
+ public void write_args(org.apache.thrift.protocol.TProtocol prot) throws org.apache.thrift.TException {
+ prot.writeMessageBegin(new org.apache.thrift.protocol.TMessage("get_part_specs_by_filter", org.apache.thrift.protocol.TMessageType.CALL, 0));
+ get_part_specs_by_filter_args args = new get_part_specs_by_filter_args();
+ args.setDb_name(db_name);
+ args.setTbl_name(tbl_name);
+ args.setFilter(filter);
+ args.setMax_parts(max_parts);
+ args.write(prot);
+ prot.writeMessageEnd();
+ }
+
+ public List<PartitionSpec> getResult() throws MetaException, NoSuchObjectException, org.apache.thrift.TException {
+ if (getState() != org.apache.thrift.async.TAsyncMethodCall.State.RESPONSE_READ) {
+ throw new IllegalStateException("Method call not finished!");
+ }
+ org.apache.thrift.transport.TMemoryInputTransport memoryTransport = new org.apache.thrift.transport.TMemoryInputTransport(getFrameBuffer().array());
+ org.apache.thrift.protocol.TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport);
+ return (new Client(prot)).recv_get_part_specs_by_filter();
+ }
+ }
+
public void get_partitions_by_expr(PartitionsByExprRequest req, org.apache.thrift.async.AsyncMethodCallback<get_partitions_by_expr_call> resultHandler) throws org.apache.thrift.TException {
checkReady();
get_partitions_by_expr_call method_call = new get_partitions_by_expr_call(req, resultHandler, this, ___protocolFactory, ___transport);
@@ -7952,6 +8170,7 @@ public class ThriftHiveMetastore {
processMap.put("add_partition", new add_partition());
processMap.put("add_partition_with_environment_context", new add_partition_with_environment_context());
processMap.put("add_partitions", new add_partitions());
+ processMap.put("add_partitions_pspec", new add_partitions_pspec());
processMap.put("append_partition", new append_partition());
processMap.put("add_partitions_req", new add_partitions_req());
processMap.put("append_partition_with_environment_context", new append_partition_with_environment_context());
@@ -7968,11 +8187,13 @@ public class ThriftHiveMetastore {
processMap.put("get_partition_by_name", new get_partition_by_name());
processMap.put("get_partitions", new get_partitions());
processMap.put("get_partitions_with_auth", new get_partitions_with_auth());
+ processMap.put("get_partitions_pspec", new get_partitions_pspec());
processMap.put("get_partition_names", new get_partition_names());
processMap.put("get_partitions_ps", new get_partitions_ps());
processMap.put("get_partitions_ps_with_auth", new get_partitions_ps_with_auth());
processMap.put("get_partition_names_ps", new get_partition_names_ps());
processMap.put("get_partitions_by_filter", new get_partitions_by_filter());
+ processMap.put("get_part_specs_by_filter", new get_part_specs_by_filter());
processMap.put("get_partitions_by_expr", new get_partitions_by_expr());
processMap.put("get_partitions_by_names", new get_partitions_by_names());
processMap.put("alter_partition", new alter_partition());
@@ -8785,6 +9006,35 @@ public class ThriftHiveMetastore {
}
}
+ public static class add_partitions_pspec<I extends Iface> extends org.apache.thrift.ProcessFunction<I, add_partitions_pspec_args> {
+ public add_partitions_pspec() {
+ super("add_partitions_pspec");
+ }
+
+ public add_partitions_pspec_args getEmptyArgsInstance() {
+ return new add_partitions_pspec_args();
+ }
+
+ protected boolean isOneway() {
+ return false;
+ }
+
+ public add_partitions_pspec_result getResult(I iface, add_partitions_pspec_args args) throws org.apache.thrift.TException {
+ add_partitions_pspec_result result = new add_partitions_pspec_result();
+ try {
+ result.success = iface.add_partitions_pspec(args.new_parts);
+ result.setSuccessIsSet(true);
+ } catch (InvalidObjectException o1) {
+ result.o1 = o1;
+ } catch (AlreadyExistsException o2) {
+ result.o2 = o2;
+ } catch (MetaException o3) {
+ result.o3 = o3;
+ }
+ return result;
+ }
+ }
+
public static class append_partition<I extends Iface> extends org.apache.thrift.ProcessFunction<I, append_partition_args> {
public append_partition() {
super("append_partition");
@@ -9219,6 +9469,32 @@ public class ThriftHiveMetastore {
}
}
+ public static class get_partitions_pspec<I extends Iface> extends org.apache.thrift.ProcessFunction<I, get_partitions_pspec_args> {
+ public get_partitions_pspec() {
+ super("get_partitions_pspec");
+ }
+
+ public get_partitions_pspec_args getEmptyArgsInstance() {
+ return new get_partitions_pspec_args();
+ }
+
+ protected boolean isOneway() {
+ return false;
+ }
+
+ public get_partitions_pspec_result getResult(I iface, get_partitions_pspec_args args) throws org.apache.thrift.TException {
+ get_partitions_pspec_result result = new get_partitions_pspec_result();
+ try {
+ result.success = iface.get_partitions_pspec(args.db_name, args.tbl_name, args.max_parts);
+ } catch (NoSuchObjectException o1) {
+ result.o1 = o1;
+ } catch (MetaException o2) {
+ result.o2 = o2;
+ }
+ return result;
+ }
+ }
+
public static class get_partition_names<I extends Iface> extends org.apache.thrift.ProcessFunction<I, get_partition_names_args> {
public get_partition_names() {
super("get_partition_names");
@@ -9347,6 +9623,32 @@ public class ThriftHiveMetastore {
}
}
+ public static class get_part_specs_by_filter<I extends Iface> extends org.apache.thrift.ProcessFunction<I, get_part_specs_by_filter_args> {
+ public get_part_specs_by_filter() {
+ super("get_part_specs_by_filter");
+ }
+
+ public get_part_specs_by_filter_args getEmptyArgsInstance() {
+ return new get_part_specs_by_filter_args();
+ }
+
+ protected boolean isOneway() {
+ return false;
+ }
+
+ public get_part_specs_by_filter_result getResult(I iface, get_part_specs_by_filter_args args) throws org.apache.thrift.TException {
+ get_part_specs_by_filter_result result = new get_part_specs_by_filter_result();
+ try {
+ result.success = iface.get_part_specs_by_filter(args.db_name, args.tbl_name, args.filter, args.max_parts);
+ } catch (MetaException o1) {
+ result.o1 = o1;
+ } catch (NoSuchObjectException o2) {
+ result.o2 = o2;
+ }
+ return result;
+ }
+ }
+
public static class get_partitions_by_expr<I extends Iface> extends org.apache.thrift.ProcessFunction<I, get_partitions_by_expr_args> {
public get_partitions_by_expr() {
super("get_partitions_by_expr");
@@ -16386,13 +16688,13 @@ public class ThriftHiveMetastore {
case 0: // SUCCESS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list458 = iprot.readListBegin();
- struct.success = new ArrayList<String>(_list458.size);
- for (int _i459 = 0; _i459 < _list458.size; ++_i459)
+ org.apache.thrift.protocol.TList _list492 = iprot.readListBegin();
+ struct.success = new ArrayList<String>(_list492.size);
+ for (int _i493 = 0; _i493 < _list492.size; ++_i493)
{
- String _elem460; // required
- _elem460 = iprot.readString();
- struct.success.add(_elem460);
+ String _elem494; // required
+ _elem494 = iprot.readString();
+ struct.success.add(_elem494);
}
iprot.readListEnd();
}
@@ -16427,9 +16729,9 @@ public class ThriftHiveMetastore {
oprot.writeFieldBegin(SUCCESS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size()));
- for (String _iter461 : struct.success)
+ for (String _iter495 : struct.success)
{
- oprot.writeString(_iter461);
+ oprot.writeString(_iter495);
}
oprot.writeListEnd();
}
@@ -16468,9 +16770,9 @@ public class ThriftHiveMetastore {
if (struct.isSetSuccess()) {
{
oprot.writeI32(struct.success.size());
- for (String _iter462 : struct.success)
+ for (String _iter496 : struct.success)
{
- oprot.writeString(_iter462);
+ oprot.writeString(_iter496);
}
}
}
@@ -16485,13 +16787,13 @@ public class ThriftHiveMetastore {
BitSet incoming = iprot.readBitSet(2);
if (incoming.get(0)) {
{
- org.apache.thrift.protocol.TList _list463 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
- struct.success = new ArrayList<String>(_list463.size);
- for (int _i464 = 0; _i464 < _list463.size; ++_i464)
+ org.apache.thrift.protocol.TList _list497 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
+ struct.success = new ArrayList<String>(_list497.size);
+ for (int _i498 = 0; _i498 < _list497.size; ++_i498)
{
- String _elem465; // required
- _elem465 = iprot.readString();
- struct.success.add(_elem465);
+ String _elem499; // required
+ _elem499 = iprot.readString();
+ struct.success.add(_elem499);
}
}
struct.setSuccessIsSet(true);
@@ -17148,13 +17450,13 @@ public class ThriftHiveMetastore {
case 0: // SUCCESS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list466 = iprot.readListBegin();
- struct.success = new ArrayList<String>(_list466.size);
- for (int _i467 = 0; _i467 < _list466.size; ++_i467)
+ org.apache.thrift.protocol.TList _list500 = iprot.readListBegin();
+ struct.success = new ArrayList<String>(_list500.size);
+ for (int _i501 = 0; _i501 < _list500.size; ++_i501)
{
- String _elem468; // required
- _elem468 = iprot.readString();
- struct.success.add(_elem468);
+ String _elem502; // required
+ _elem502 = iprot.readString();
+ struct.success.add(_elem502);
}
iprot.readListEnd();
}
@@ -17189,9 +17491,9 @@ public class ThriftHiveMetastore {
oprot.writeFieldBegin(SUCCESS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size()));
- for (String _iter469 : struct.success)
+ for (String _iter503 : struct.success)
{
- oprot.writeString(_iter469);
+ oprot.writeString(_iter503);
}
oprot.writeListEnd();
}
@@ -17230,9 +17532,9 @@ public class ThriftHiveMetastore {
if (struct.isSetSuccess()) {
{
oprot.writeI32(struct.success.size());
- for (String _iter470 : struct.success)
+ for (String _iter504 : struct.success)
{
- oprot.writeString(_iter470);
+ oprot.writeString(_iter504);
}
}
}
@@ -17247,13 +17549,13 @@ public class ThriftHiveMetastore {
BitSet incoming = iprot.readBitSet(2);
if (incoming.get(0)) {
{
- org.apache.thrift.protocol.TList _list471 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
- struct.success = new ArrayList<String>(_list471.size);
- for (int _i472 = 0; _i472 < _list471.size; ++_i472)
+ org.apache.thrift.protocol.TList _list505 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
+ struct.success = new ArrayList<String>(_list505.size);
+ for (int _i506 = 0; _i506 < _list505.size; ++_i506)
{
- String _elem473; // required
- _elem473 = iprot.readString();
- struct.success.add(_elem473);
+ String _elem507; // required
+ _elem507 = iprot.readString();
+ struct.success.add(_elem507);
}
}
struct.setSuccessIsSet(true);
@@ -21860,16 +22162,16 @@ public class ThriftHiveMetastore {
case 0: // SUCCESS
if (schemeField.type == org.apache.thrift.protocol.TType.MAP) {
{
- org.apache.thrift.protocol.TMap _map474 = iprot.readMapBegin();
- struct.success = new HashMap<String,Type>(2*_map474.size);
- for (int _i475 = 0; _i475 < _map474.size; ++_i475)
+ org.apache.thrift.protocol.TMap _map508 = iprot.readMapBegin();
+ struct.success = new HashMap<String,Type>(2*_map508.size);
+ for (int _i509 = 0; _i509 < _map508.size; ++_i509)
{
- String _key476; // required
- Type _val477; // required
- _key476 = iprot.readString();
- _val477 = new Type();
- _val477.read(iprot);
- struct.success.put(_key476, _val477);
+ String _key510; // required
+ Type _val511; // required
+ _key510 = iprot.readString();
+ _val511 = new Type();
+ _val511.read(iprot);
+ struct.success.put(_key510, _val511);
}
iprot.readMapEnd();
}
@@ -21904,10 +22206,10 @@ public class ThriftHiveMetastore {
oprot.writeFieldBegin(SUCCESS_FIELD_DESC);
{
oprot.writeMapBegin(new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRUCT, struct.success.size()));
- for (Map.Entry<String, Type> _iter478 : struct.success.entrySet())
+ for (Map.Entry<String, Type> _iter512 : struct.success.entrySet())
{
- oprot.writeString(_iter478.getKey());
- _iter478.getValue().write(oprot);
+ oprot.writeString(_iter512.getKey());
+ _iter512.getValue().write(oprot);
}
oprot.writeMapEnd();
}
@@ -21946,10 +22248,10 @@ public class ThriftHiveMetastore {
if (struct.isSetSuccess()) {
{
oprot.writeI32(struct.success.size());
- for (Map.Entry<String, Type> _iter479 : struct.success.entrySet())
+ for (Map.Entry<String, Type> _iter513 : struct.success.entrySet())
{
- oprot.writeString(_iter479.getKey());
- _iter479.getValue().write(oprot);
+ oprot.writeString(_iter513.getKey());
+ _iter513.getValue().write(oprot);
}
}
}
@@ -21964,16 +22266,16 @@ public class ThriftHiveMetastore {
BitSet incoming = iprot.readBitSet(2);
if (incoming.get(0)) {
{
- org.apache.thrift.protocol.TMap _map480 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
- struct.success = new HashMap<String,Type>(2*_map480.size);
- for (int _i481 = 0; _i481 < _map480.size; ++_i481)
+ org.apache.thrift.protocol.TMap _map514 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
+ struct.success = new HashMap<String,Type>(2*_map514.size);
+ for (int _i515 = 0; _i515 < _map514.size; ++_i515)
{
- String _key482; // required
- Type _val483; // required
- _key482 = iprot.readString();
- _val483 = new Type();
- _val483.read(iprot);
- struct.success.put(_key482, _val483);
+ String _key516; // required
+ Type _val517; // required
+ _key516 = iprot.readString();
+ _val517 = new Type();
+ _val517.read(iprot);
+ struct.success.put(_key516, _val517);
}
}
struct.setSuccessIsSet(true);
@@ -23008,14 +23310,14 @@ public class ThriftHiveMetastore {
case 0: // SUCCESS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list484 = iprot.readListBegin();
- struct.success = new ArrayList<FieldSchema>(_list484.size);
- for (int _i485 = 0; _i485 < _list484.size; ++_i485)
+ org.apache.thrift.protocol.TList _list518 = iprot.readListBegin();
+ struct.success = new ArrayList<FieldSchema>(_list518.size);
+ for (int _i519 = 0; _i519 < _list518.size; ++_i519)
{
- FieldSchema _elem486; // required
- _elem486 = new FieldSchema();
- _elem486.read(iprot);
- struct.success.add(_elem486);
+ FieldSchema _elem520; // required
+ _elem520 = new FieldSchema();
+ _elem520.read(iprot);
+ struct.success.add(_elem520);
}
iprot.readListEnd();
}
@@ -23068,9 +23370,9 @@ public class ThriftHiveMetastore {
oprot.writeFieldBegin(SUCCESS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size()));
- for (FieldSchema _iter487 : struct.success)
+ for (FieldSchema _iter521 : struct.success)
{
- _iter487.write(oprot);
+ _iter521.write(oprot);
}
oprot.writeListEnd();
}
@@ -23125,9 +23427,9 @@ public class ThriftHiveMetastore {
if (struct.isSetSuccess()) {
{
oprot.writeI32(struct.success.size());
- for (FieldSchema _iter488 : struct.success)
+ for (FieldSchema _iter522 : struct.success)
{
- _iter488.write(oprot);
+ _iter522.write(oprot);
}
}
}
@@ -23148,14 +23450,14 @@ public class ThriftHiveMetastore {
BitSet incoming = iprot.readBitSet(4);
if (incoming.get(0)) {
{
- org.apache.thrift.protocol.TList _list489 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
- struct.success = new ArrayList<FieldSchema>(_list489.size);
- for (int _i490 = 0; _i490 < _list489.size; ++_i490)
+ org.apache.thrift.protocol.TList _list523 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
+ struct.success = new ArrayList<FieldSchema>(_list523.size);
+ for (int _i524 = 0; _i524 < _list523.size; ++_i524)
{
- FieldSchema _elem491; // required
- _elem491 = new FieldSchema();
- _elem491.read(iprot);
- struct.success.add(_elem491);
+ FieldSchema _elem525; // required
+ _elem525 = new FieldSchema();
+ _elem525.read(iprot);
+ struct.success.add(_elem525);
}
}
struct.setSuccessIsSet(true);
@@ -24200,14 +24502,14 @@ public class ThriftHiveMetastore {
case 0: // SUCCESS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list492 = iprot.readListBegin();
- struct.success = new ArrayList<FieldSchema>(_list492.size);
- for (int _i493 = 0; _i493 < _list492.size; ++_i493)
+ org.apache.thrift.protocol.TList _list526 = iprot.readListBegin();
+ struct.success = new ArrayList<FieldSchema>(_list526.size);
+ for (int _i527 = 0; _i527 < _list526.size; ++_i527)
{
- FieldSchema _elem494; // required
- _elem494 = new FieldSchema();
- _elem494.read(iprot);
- struct.success.add(_elem494);
+ FieldSchema _elem528; // required
+ _elem528 = new FieldSchema();
+ _elem528.read(iprot);
+ struct.success.add(_elem528);
}
iprot.readListEnd();
}
@@ -24260,9 +24562,9 @@ public class ThriftHiveMetastore {
oprot.writeFieldBegin(SUCCESS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size()));
- for (FieldSchema _iter495 : struct.success)
+ for (FieldSchema _iter529 : struct.success)
{
- _iter495.write(oprot);
+ _iter529.write(oprot);
}
oprot.writeListEnd();
}
@@ -24317,9 +24619,9 @@ public class ThriftHiveMetastore {
if (struct.isSetSuccess()) {
{
oprot.writeI32(struct.success.size());
- for (FieldSchema _iter496 : struct.success)
+ for (FieldSchema _iter530 : struct.success)
{
- _iter496.write(oprot);
+ _iter530.write(oprot);
}
}
}
@@ -24340,14 +24642,14 @@ public class ThriftHiveMetastore {
BitSet incoming = iprot.readBitSet(4);
if (incoming.get(0)) {
{
- org.apache.thrift.protocol.TList _list497 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
- struct.success = new ArrayList<FieldSchema>(_list497.size);
- for (int _i498 = 0; _i498 < _list497.size; ++_i498)
+ org.apache.thrift.protocol.TList _list531 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
+ struct.success = new ArrayList<FieldSchema>(_list531.size);
+ for (int _i532 = 0; _i532 < _list531.size; ++_i532)
{
- FieldSchema _elem499; // required
- _elem499 = new FieldSchema();
- _elem499.read(iprot);
- struct.success.add(_elem499);
+ FieldSchema _elem533; // required
+ _elem533 = new FieldSchema();
+ _elem533.read(iprot);
+ struct.success.add(_elem533);
}
}
struct.setSuccessIsSet(true);
@@ -29590,13 +29892,13 @@ public class ThriftHiveMetastore {
case 0: // SUCCESS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list500 = iprot.readListBegin();
- struct.success = new ArrayList<String>(_list500.size);
- for (int _i501 = 0; _i501 < _list500.size; ++_i501)
+ org.apache.thrift.protocol.TList _list534 = iprot.readListBegin();
+ struct.success = new ArrayList<String>(_list534.size);
+ for (int _i535 = 0; _i535 < _list534.size; ++_i535)
{
- String _elem502; // required
- _elem502 = iprot.readString();
- struct.success.add(_elem502);
+ String _elem536; // required
+ _elem536 = iprot.readString();
+ struct.success.add(_elem536);
}
iprot.readListEnd();
}
@@ -29631,9 +29933,9 @@ public class ThriftHiveMetastore {
oprot.writeFieldBegin(SUCCESS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size()));
- for (String _iter503 : struct.success)
+ for (String _iter537 : struct.success)
{
- oprot.writeString(_iter503);
+ oprot.writeString(_iter537);
}
oprot.writeListEnd();
}
@@ -29672,9 +29974,9 @@ public class ThriftHiveMetastore {
if (struct.isSetSuccess()) {
{
oprot.writeI32(struct.success.size());
- for (String _iter504 : struct.success)
+ for (String _iter538 : struct.success)
{
- oprot.writeString(_iter504);
+ oprot.writeString(_iter538);
}
}
}
@@ -29689,13 +29991,13 @@ public class ThriftHiveMetastore {
BitSet incoming = iprot.readBitSet(2);
if (incoming.get(0)) {
{
- org.apache.thrift.protocol.TList _list505 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
- struct.success = new ArrayList<String>(_list505.size);
- for (int _i506 = 0; _i506 < _list505.size; ++_i506)
+ org.apache.thrift.protocol.TList _list539 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
+ struct.success = new ArrayList<String>(_list539.size);
+ for (int _i540 = 0; _i540 < _list539.size; ++_i540)
{
- String _elem507; // required
- _elem507 = iprot.readString();
- struct.success.add(_elem507);
+ String _elem541; // required
+ _elem541 = iprot.readString();
+ struct.success.add(_elem541);
}
}
struct.setSuccessIsSet(true);
@@ -30464,13 +30766,13 @@ public class ThriftHiveMetastore {
case 0: // SUCCESS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list508 = iprot.readListBegin();
- struct.success = new ArrayList<String>(_list508.size);
- for (int _i509 = 0; _i509 < _list508.size; ++_i509)
+ org.apache.thrift.protocol.TList _list542 = iprot.readListBegin();
+ struct.success = new ArrayList<String>(_list542.size);
+ for (int _i543 = 0; _i543 < _list542.size; ++_i543)
{
- String _elem510; // required
- _elem510 = iprot.readString();
- struct.success.add(_elem510);
+ String _elem544; // required
+ _elem544 = iprot.readString();
+ struct.success.add(_elem544);
}
iprot.readListEnd();
}
@@ -30505,9 +30807,9 @@ public class ThriftHiveMetastore {
oprot.writeFieldBegin(SUCCESS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size()));
- for (String _iter511 : struct.success)
+ for (String _iter545 : struct.success)
{
- oprot.writeString(_iter511);
+ oprot.writeString(_iter545);
}
oprot.writeListEnd();
}
@@ -30546,9 +30848,9 @@ public class ThriftHiveMetastore {
if (struct.isSetSuccess()) {
{
oprot.writeI32(struct.success.size());
- for (String _iter512 : struct.success)
+ for (String _iter546 : struct.success)
{
- oprot.writeString(_iter512);
+ oprot.writeString(_iter546);
}
}
}
@@ -30563,13 +30865,13 @@ public class ThriftHiveMetastore {
BitSet incoming = iprot.readBitSet(2);
if (incoming.get(0)) {
{
- org.apache.thrift.protocol.TList _list513 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
- struct.success = new ArrayList<String>(_list513.size);
- for (int _i514 = 0; _i514 < _list513.size; ++_i514)
+ org.apache.thrift.protocol.TList _list547 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
+ struct.success = new ArrayList<String>(_list547.size);
+ for (int _i548 = 0; _i548 < _list547.size; ++_i548)
{
- String _elem515; // required
- _elem515 = iprot.readString();
- struct.success.add(_elem515);
+ String _elem549; // required
+ _elem549 = iprot.readString();
+ struct.success.add(_elem549);
}
}
struct.setSuccessIsSet(true);
@@ -32025,13 +32327,13 @@ public class ThriftHiveMetastore {
case 2: // TBL_NAMES
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list516 = iprot.readListBegin();
- struct.tbl_names = new ArrayList<String>(_list516.size);
- for (int _i517 = 0; _i517 < _list516.size; ++_i517)
+ org.apache.thrift.protocol.TList _list550 = iprot.readListBegin();
+ struct.tbl_names = new ArrayList<String>(_list550.size);
+ for (int _i551 = 0; _i551 < _list550.size; ++_i551)
{
- String _elem518; // required
- _elem518 = iprot.readString();
- struct.tbl_names.add(_elem518);
+ String _elem552; // required
+ _elem552 = iprot.readString();
+ struct.tbl_names.add(_elem552);
}
iprot.readListEnd();
}
@@ -32062,9 +32364,9 @@ public class ThriftHiveMetastore {
oprot.writeFieldBegin(TBL_NAMES_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.tbl_names.size()));
- for (String _iter519 : struct.tbl_names)
+ for (String _iter553 : struct.tbl_names)
{
- oprot.writeString(_iter519);
+ oprot.writeString(_iter553);
}
oprot.writeListEnd();
}
@@ -32101,9 +32403,9 @@ public class ThriftHiveMetastore {
if (struct.isSetTbl_names()) {
{
oprot.writeI32(struct.tbl_names.size());
- for (String _iter520 : struct.tbl_names)
+ for (String _iter554 : struct.tbl_names)
{
- oprot.writeString(_iter520);
+ oprot.writeString(_iter554);
}
}
}
@@ -32119,13 +32421,13 @@ public class ThriftHiveMetastore {
}
if (incoming.get(1)) {
{
- org.apache.thrift.protocol.TList _list521 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
- struct.tbl_names = new ArrayList<String>(_list521.size);
- for (int _i522 = 0; _i522 < _list521.size; ++_i522)
+ org.apache.thrift.protocol.TList _list555 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
+ struct.tbl_names = new ArrayList<String>(_list555.size);
+ for (int _i556 = 0; _i556 < _list555.size; ++_i556)
{
- String _elem523; // required
- _elem523 = iprot.readString();
- struct.tbl_names.add(_elem523);
+ String _elem557; // required
+ _elem557 = iprot.readString();
+ struct.tbl_names.add(_elem557);
}
}
struct.setTbl_namesIsSet(true);
@@ -32693,14 +32995,14 @@ public class ThriftHiveMetastore {
case 0: // SUCCESS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list524 = iprot.readListBegin();
- struct.success = new ArrayList<Table>(_list524.size);
- for (int _i525 = 0; _i525 < _list524.size; ++_i525)
+ org.apache.thrift.protocol.TList _list558 = iprot.readListBegin();
+ struct.success = new ArrayList<Table>(_list558.size);
+ for (int _i559 = 0; _i559 < _list558.size; ++_i559)
{
- Table _elem526; // required
- _elem526 = new Table();
- _elem526.read(iprot);
- struct.success.add(_elem526);
+ Table _elem560; // required
+ _elem560 = new Table();
+ _elem560.read(iprot);
+ struct.success.add(_elem560);
}
iprot.readListEnd();
}
@@ -32753,9 +33055,9 @@ public class ThriftHiveMetastore {
oprot.writeFieldBegin(SUCCESS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size()));
- for (Table _iter527 : struct.success)
+ for (Table _iter561 : struct.success)
{
- _iter527.write(oprot);
+ _iter561.write(oprot);
}
oprot.writeListEnd();
}
@@ -32810,9 +33112,9 @@ public class ThriftHiveMetastore {
if (struct.isSetSuccess()) {
{
oprot.writeI32(struct.success.size());
- for (Table _iter528 : struct.success)
+ for (Table _iter562 : struct.success)
{
- _iter528.write(oprot);
+ _iter562.write(oprot);
}
}
}
@@ -32833,14 +33135,14 @@ public class ThriftHiveMetastore {
BitSet incoming = iprot.readBitSet(4);
if (incoming.get(0)) {
{
- org.apache.thrift.protocol.TList _list529 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
- struct.success = new ArrayList<Table>(_list529.size);
- for (int _i530 = 0; _i530 < _list529.size; ++_i530)
+ org.apache.thrift.protocol.TList _list563 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
+ struct.success = new ArrayList<Table>(_list563.size);
+ for (int _i564 = 0; _i564 < _list563.size; ++_i564)
{
- Table _elem531; // required
- _elem531 = new Table();
- _elem531.read(iprot);
- struct.success.add(_elem531);
+ Table _elem565; // required
+ _elem565 = new Table();
+ _elem565.read(iprot);
+ struct.success.add(_elem565);
}
}
struct.setSuccessIsSet(true);
@@ -33989,13 +34291,13 @@ public class ThriftHiveMetastore {
case 0: // SUCCESS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list532 = iprot.readListBegin();
- struct.success = new ArrayList<String>(_list532.size);
- for (int _i533 = 0; _i533 < _list532.size; ++_i533)
+ org.apache.thrift.protocol.TList _list566 = iprot.readListBegin();
+ struct.success = new ArrayList<String>(_list566.size);
+ for (int _i567 = 0; _i567 < _list566.size; ++_i567)
{
- String _elem534; // required
- _elem534 = iprot.readString();
- struct.success.add(_elem534);
+ String _elem568; // required
+ _elem568 = iprot.readString();
+ struct.success.add(_elem568);
}
iprot.readListEnd();
}
@@ -34048,9 +34350,9 @@ public class ThriftHiveMetastore {
oprot.writeFieldBegin(SUCCESS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size()));
- for (String _iter535 : struct.success)
+ for (String _iter569 : struct.success)
{
- oprot.writeString(_iter535);
+ oprot.writeString(_iter569);
}
oprot.writeListEnd();
}
@@ -34105,9 +34407,9 @@ public class ThriftHiveMetastore {
if (struct.isSetSuccess()) {
{
oprot.writeI32(struct.success.size());
- for (String _iter536 : struct.success)
+ for (String _iter570 : struct.success)
{
- oprot.writeString(_iter536);
+ oprot.writeString(_iter570);
}
}
}
@@ -34128,13 +34430,13 @@ public class ThriftHiveMetastore {
BitSet incoming = iprot.readBitSet(4);
if (incoming.get(0)) {
{
- org.apache.thrift.protocol.TList _list537 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
- struct.success = new ArrayList<String>(_list537.size);
- for (int _i538 = 0; _i538 < _list537.size; ++_i538)
+ org.apache.thrift.protocol.TList _list571 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
+ struct.success = new ArrayList<String>(_list571.size);
+ for (int _i572 = 0; _i572 < _list571.size; ++_i572)
{
- String _elem539; // required
- _elem539 = iprot.readString();
- struct.success.add(_elem539);
+ String _elem573; // required
+ _elem573 = iprot.readString();
+ struct.success.add(_elem573);
}
}
struct.setSuccessIsSet(true);
@@ -38854,14 +39156,14 @@ public class ThriftHiveMetastore {
case 1: // NEW_PARTS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list540 = iprot.readListBegin();
- struct.new_parts = new ArrayList<Partition>(_list540.size);
- for (int _i541 = 0; _i541 < _list540.size; ++_i541)
+ org.apache.thrift.protocol.TList _list574 = iprot.readListBegin();
+ struct.new_parts = new ArrayList<Partition>(_list574.size);
+ for (int _i575 = 0; _i575 < _list574.size; ++_i575)
{
- Partition _elem542; // required
- _elem542 = new Partition();
- _elem542.read(iprot);
- struct.new_parts.add(_elem542);
+ Partition _elem576; // required
+ _elem576 = new Partition();
+ _elem576.read(iprot);
+ struct.new_parts.add(_elem576);
}
iprot.readListEnd();
}
@@ -38887,9 +39189,9 @@ public class ThriftHiveMetastore {
oprot.writeFieldBegin(NEW_PARTS_FIELD_DESC);
{
oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.new_parts.size()));
- for (Partition _iter543 : struct.new_parts)
+ for (Partition _iter577 : struct.new_parts)
{
- _iter543.write(oprot);
+ _iter577.write(oprot);
}
oprot.writeListEnd();
}
@@ -38920,9 +39222,9 @@ public class ThriftHiveMetastore {
if (struct.isSetNew_parts()) {
{
oprot.writeI32(struct.new_parts.size());
- for (Partition _iter544 : struct.new_parts)
+ for (Partition _iter578 : struct.new_parts)
{
- _iter544.write(oprot);
+ _iter578.write(oprot);
}
}
}
@@ -38934,14 +39236,14 @@ public class ThriftHiveMetastore {
BitSet incoming = iprot.readBitSet(1);
if (incoming.get(0)) {
{
- org.apache.thrift.protocol.TList _list545 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
- struct.new_parts = new ArrayList<Partition>(_list545.size);
- for (int _i546 = 0; _i546 < _list545.size; ++_i546)
+ org.apache.thrift.protocol.TList _list579 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
+ struct.new_parts = new ArrayList<Partition>(_list579.size);
+ for (int _i580 = 0; _i580 < _list579.size; ++_i580)
{
- Partition _elem547; // required
- _elem547 = new Partition();
- _elem547.read(iprot);
- struct.new_parts.add(_elem547);
+ Partition _elem581; // required
+ _elem581 = new Partition();
+ _elem581.read(iprot);
+ struct.new_parts.add(_elem581);
}
}
struct.setNew_partsIsSet(true);
@@ -39627,28 +39929,22 @@ public class ThriftHiveMetastore {
}
- public static class append_partition_args implements org.apache.thrift.TBase<append_partition_args, append_partition_args._Fields>, java.io.Serializable, Cloneable {
- private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("append_partition_args");
+ public static class add_partitions_pspec_args implements org.apache.thrift.TBase<add_partitions_pspec_args, add_partitions_pspec_args._Fields>, java.io.Serializable, Cloneable {
+ private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("add_partitions_pspec_args");
- private static final org.apache.thrift.protocol.TField DB_NAME_FIELD_DESC = new org.apache.thrift.protocol.TField("db_name", org.apache.thrift.protocol.TType.STRING, (short)1);
- private static final org.apache.thrift.protocol.TField TBL_NAME_FIELD_DESC = new org.apache.thrift.protocol.TField("tbl_name", org.apache.thrift.protocol.TType.STRING, (short)2);
- private static final org.apache.thrift.protocol.TField PART_VALS_FIELD_DESC = new org.apache.thrift.protocol.TField("part_vals", org.apache.thrift.protocol.TType.LIST, (short)3);
+ private static final org.apache.thrift.protocol.TField NEW_PARTS_FIELD_DESC = new org.apache.thrift.protocol.TField("new_parts", org.apache.thrift.protocol.TType.LIST, (short)1);
private static final Map<Class<? extends IScheme>, SchemeFactory> schemes = new HashMap<Class<? extends IScheme>, SchemeFactory>();
static {
- schemes.put(StandardScheme.class, new append_partition_argsStandardSchemeFactory());
- schemes.put(TupleScheme.class, new append_partition_argsTupleSchemeFactory());
+ schemes.put(StandardScheme.class, new add_partitions_pspec_argsStandardSchemeFactory());
+ schemes.put(TupleScheme.class, new add_partitions_pspec_argsTupleSchemeFactory());
}
- private String db_name; // required
- private String tbl_name; // required
- private List<String> part_vals; // required
+ private List<PartitionSpec> new_parts; // required
/** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */
public enum _Fields implements org.apache.thrift.TFieldIdEnum {
- DB_NAME((short)1, "db_name"),
- TBL_NAME((short)2, "tbl_name"),
- PART_VALS((short)3, "part_vals");
+ NEW_PARTS((short)1, "new_parts");
private static final Map<String, _Fields> byName = new HashMap<String, _Fields>();
@@ -39663,12 +39959,8 @@ public class ThriftHiveMetastore {
*/
public static _Fields findByThriftId(int fieldId) {
switch(fieldId) {
- case 1: // DB_NAME
- return DB_NAME;
- case 2: // TBL_NAME
- return TBL_NAME;
- case 3: // PART_VALS
- return PART_VALS;
+ case 1: // NEW_PARTS
+ return NEW_PARTS;
default:
return null;
}
@@ -39712,168 +40004,90 @@ public class ThriftHiveMetastore {
public static final Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap;
static {
Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class);
- tmpMap.put(_Fields.DB_NAME, new org.apache.thrift.meta_data.FieldMetaData("db_name", org.apache.thrift.TFieldRequirementType.DEFAULT,
- new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING)));
- tmpMap.put(_Fields.TBL_NAME, new org.apache.thrift.meta_data.FieldMetaData("tbl_name", org.apache.thrift.TFieldRequirementType.DEFAULT,
- new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING)));
- tmpMap.put(_Fields.PART_VALS, new org.apache.thrift.meta_data.FieldMetaData("part_vals", org.apache.thrift.TFieldRequirementType.DEFAULT,
+ tmpMap.put(_Fields.NEW_PARTS, new org.apache.thrift.meta_data.FieldMetaData("new_parts", org.apache.thrift.TFieldRequirementType.DEFAULT,
new org.apache.thrift.meta_data.ListMetaData(org.apache.thrift.protocol.TType.LIST,
- new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))));
+ new org.apache.thrift.meta_data.StructMetaData(org.apache.thrift.protocol.TType.STRUCT, PartitionSpec.class))));
metaDataMap = Collections.unmodifiableMap(tmpMap);
- org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(append_partition_args.class, metaDataMap);
+ org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(add_partitions_pspec_args.class, metaDataMap);
}
- public append_partition_args() {
+ public add_partitions_pspec_args() {
}
- public append_partition_args(
- String db_name,
- String tbl_name,
- List<String> part_vals)
+ public add_partitions_pspec_args(
+ List<PartitionSpec> new_parts)
{
this();
- this.db_name = db_name;
- this.tbl_name = tbl_name;
- this.part_vals = part_vals;
+ this.new_parts = new_parts;
}
/**
* Performs a deep copy on <i>other</i>.
*/
- public append_partition_args(append_partition_args other) {
- if (other.isSetDb_name()) {
- this.db_name = other.db_name;
- }
- if (other.isSetTbl_name()) {
- this.tbl_name = other.tbl_name;
- }
- if (other.isSetPart_vals()) {
- List<String> __this__part_vals = new ArrayList<String>();
- for (String other_element : other.part_vals) {
- __this__part_vals.add(other_element);
+ public add_partitions_pspec_args(add_partitions_pspec_args other) {
+ if (other.isSetNew_parts()) {
+ List<PartitionSpec> __this__new_parts = new ArrayList<PartitionSpec>();
+ for (PartitionSpec other_element : other.new_parts) {
+ __this__new_parts.add(new PartitionSpec(other_element));
}
- this.part_vals = __this__part_vals;
+ this.new_parts = __this__new_parts;
}
}
- public append_partition_args deepCopy() {
- return new append_partition_args(this);
+ public add_partitions_pspec_args deepCopy() {
+ return new add_partitions_pspec_args(this);
}
@Override
public void clear() {
- this.db_name = null;
- this.tbl_name = null;
- this.part_vals = null;
- }
-
- public String getDb_name() {
- return this.db_name;
- }
-
- public void setDb_name(String db_name) {
- this.db_name = db_name;
- }
-
- public void unsetDb_name() {
- this.db_name = null;
- }
-
- /** Returns true if field db_name is set (has been assigned a value) and false otherwise */
- public boolean isSetDb_name() {
- return this.db_name != null;
- }
-
- public void setDb_nameIsSet(boolean value) {
- if (!value) {
- this.db_name = null;
- }
- }
-
- public String getTbl_name() {
- return this.tbl_name;
- }
-
- public void setTbl_name(String tbl_name) {
- this.tbl_name = tbl_name;
- }
-
- public void unsetTbl_name() {
- this.tbl_name = null;
- }
-
- /** Returns true if field tbl_name is set (has been assigned a value) and false otherwise */
- public boolean isSetTbl_name() {
- return this.tbl_name != null;
- }
-
- public void setTbl_nameIsSet(boolean value) {
- if (!value) {
- this.tbl_name = null;
- }
+ this.new_parts = null;
}
- public int getPart_valsSize() {
- return (this.part_vals == null) ? 0 : this.part_vals.size();
+ public int getNew_partsSize() {
+ return (this.new_parts == null) ? 0 : this.new_parts.size();
}
- public java.util.Iterator<String> getPart_valsIterator() {
- return (this.part_vals == null) ? null : this.part_vals.iterator();
+ public java.util.Iterator<PartitionSpec> getNew_partsIterator() {
+ return (this.new_parts == null) ? null : this.new_parts.iterator();
}
- public void addToPart_vals(String elem) {
- if (this.part_vals == null) {
- this.part_vals = new ArrayList<String>();
+ public void addToNew_parts(PartitionSpec elem) {
+ if (this.new_parts == null) {
+ this.new_parts = new ArrayList<PartitionSpec>();
}
- this.part_vals.add(elem);
+ this.new_parts.add(elem);
}
- public List<String> getPart_vals() {
- return this.part_vals;
+ public List<PartitionSpec> getNew_parts() {
+ return this.new_parts;
}
- public void setPart_vals(List<String> part_vals) {
- this.part_vals = part_vals;
+ public void setNew_parts(List<PartitionSpec> new_parts) {
+ this.new_parts = new_parts;
}
- public void unsetPart_vals() {
- this.part_vals = null;
+ public void unsetNew_parts() {
+ this.new_parts = null;
}
- /** Returns true if field part_vals is set (has been assigned a value) and false otherwise */
- public boolean isSetPart_vals() {
- return this.part_vals != null;
+ /** Returns true if field new_parts is set (has been assigned a value) and false otherwise */
+ public boolean isSetNew_parts() {
+ return this.new_parts != null;
}
- public void setPart_valsIsSet(boolean value) {
+ public void setNew_partsIsSet(boolean value) {
if (!value) {
- this.part_vals = null;
+ this.new_parts = null;
}
}
public void setFieldValue(_Fields field, Object value) {
switch (field) {
- case DB_NAME:
- if (value == null) {
- unsetDb_name();
- } else {
- setDb_name((String)value);
- }
- break;
-
- case TBL_NAME:
- if (value == null) {
- unsetTbl_name();
- } else {
- setTbl_name((String)value);
- }
- break;
-
- case PART_VALS:
+ case NEW_PARTS:
if (value == null) {
- unsetPart_vals();
+ unsetNew_parts();
} else {
- setPart_vals((List<String>)value);
+ setNew_parts((List<PartitionSpec>)value);
}
break;
@@ -39882,14 +40096,8 @@ public class ThriftHiveMetastore {
public Object getFieldValue(_Fields field) {
switch (field) {
- case DB_NAME:
- return getDb_name();
-
- case TBL_NAME:
- return getTbl_name();
-
- case PART_VALS:
- return getPart_vals();
+ case NEW_PARTS:
+ return getNew_parts();
}
throw new IllegalStateException();
@@ -39902,12 +40110,8 @@ public class ThriftHiveMetastore {
}
switch (field) {
- case DB_NAME:
- return isSetDb_name();
- case TBL_NAME:
- return isSetTbl_name();
- case PART_VALS:
- return isSetPart_vals();
+ case NEW_PARTS:
+ return isSetNew_parts();
}
throw new IllegalStateException();
}
@@ -39916,39 +40120,21 @@ public class ThriftHiveMetastore {
public boolean equals(Object that) {
if (that == null)
return false;
- if (that instanceof append_partition_args)
- return this.equals((append_partition_args)that);
+ if (that instanceof add_partitions_pspec_args)
+ return this.equals((add_partitions_pspec_args)that);
return false;
}
- public boolean equals(append_partition_args that) {
+ public boolean equals(add_partitions_pspec_args that) {
if (that == null)
return false;
- boolean this_present_db_name = true && this.isSetDb_name();
- boolean that_present_db_name = true && that.isSetDb_name();
- if (this_present_db_name || that_present_db_name) {
- if (!(this_present_db_name && that_present_db_name))
- return false;
- if (!this.db_name.equals(that.db_name))
- return false;
- }
-
- boolean this_present_tbl_name = true && this.isSetTbl_name();
- boolean that_present_tbl_name = true && that.isSetTbl_name();
- if (this_present_tbl_name || that_present_tbl_name) {
- if (!(this_present_tbl_name && that_present_tbl_name))
- return false;
- if (!this.tbl_name.equals(that.tbl_name))
- return false;
- }
-
- boolean this_present_part_vals = true && this.isSetPart_vals();
- boolean that_present_part_vals = true && that.isSetPart_vals();
- if (this_present_part_vals || that_present_part_vals) {
- if (!(this_present_part_vals && that_present_part_vals))
+ boolean this_present_new_parts = true && this.isSetNew_parts();
+ boolean that_present_new_parts = true && that.isSetNew_parts();
+ if (this_present_new_parts || that_present_new_parts) {
+ if (!(this_present_new_parts && that_present_new_parts))
return false;
- if (!this.part_vals.equals(that.part_vals))
+ if (!this.new_parts.equals(that.new_parts))
return false;
}
@@ -39959,58 +40145,28 @@ public class ThriftHiveMetastore {
public int hashCode() {
HashCodeBuilder builder = new HashCodeBuilder();
- boolean present_db_name = true && (isSetDb_name());
- builder.append(present_db_name);
- if (present_db_name)
- builder.append(db_name);
-
- boolean present_tbl_name = true && (isSetTbl_name());
- builder.append(present_tbl_name);
- if (present_tbl_name)
- builder.append(tbl_name);
-
- boolean present_part_vals = true && (isSetPart_vals());
- builder.append(present_part_vals);
- if (present_part_vals)
- builder.append(part_vals);
+ boolean present_new_parts = true && (isSetNew_parts());
+ builder.append(present_new_parts);
+ if (present_new_parts)
+ builder.append(new_parts);
return builder.toHashCode();
}
- public int compareTo(append_partition_args other) {
+ public int compareTo(add_partitions_pspec_args other) {
if (!getClass().equals(other.getClass())) {
return getClass().getName().compareTo(other.getClass().getName());
}
int lastComparison = 0;
- append_partition_args typedOther = (append_partition_args)other;
+ add_partitions_pspec_args typedOther = (add_partitions_pspec_args)other;
- lastComparison = Boolean.valueOf(isSetDb_name()).compareTo(typedOther.isSetDb_name());
- if (lastComparison != 0) {
- return lastComparison;
- }
- if (isSetDb_name()) {
- lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.db_name, typedOther.db_name);
- if (lastComparison != 0) {
- return lastComparison;
- }
- }
- lastComparison = Boolean.valueOf(isSetTbl_name()).compareTo(typedOther.isSetTbl_name());
- if (lastComparison != 0) {
- return lastComparison;
- }
- if (isSetTbl_name()) {
- lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.tbl_name, typedOther.tbl_name);
- if (lastComparison != 0) {
- return lastComparison;
- }
- }
- lastComparison = Boolean.valueOf(isSetPart_vals()).compareTo(typedOther.isSetPart_vals());
+ lastComparison = Boolean.valueOf(isSetNew_parts()).compareTo(typedOther.isSetNew_parts());
if (lastComparison != 0) {
return lastComparison;
}
- if (isSetPart_vals()) {
- lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.part_vals, typedOther.part_vals);
+ if (isSetNew_parts()) {
+ lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.new_parts, typedOther.new_parts);
if (lastComparison != 0) {
return lastComparison;
}
@@ -40032,30 +40188,14 @@ public class ThriftHiveMetastore {
@Override
public String toString() {
- StringBuilder sb = new StringBuilder("append_partition_args(");
+ StringBuilder sb = new StringBuilder("add_partitions_pspec_args(");
boolean first = true;
- sb.append("db_name:");
- if (this.db_name == null) {
- sb.append("null");
- } else {
- sb.append(this.db_name);
- }
- first = false;
- if (!first) sb.append(", ");
- sb.append("tbl_name:");
- if (this.tbl_name == null) {
- sb.append("null");
- } else {
- sb.append(this.tbl_name);
- }
- first = false;
- if (!first) sb.append(", ");
- sb.append("part_vals:");
- if (this.part_vals == null) {
+ sb.append("new_parts:");
+ if (this.new_parts == null) {
sb.append("null");
} else {
- sb.append(this.part_vals);
+ sb.append(this.new_parts);
}
first = false;
sb.append(")");
@@ -40083,15 +40223,15 @@ public class ThriftHiveMetastore {
}
}
- private static class append_partition_argsStandardSchemeFactory implements SchemeFactory {
- public append_partition_argsStandardScheme getScheme() {
- return new append_partition_argsStandardScheme();
+ private static class add_partitions_pspec_argsStandardSchemeFactory implements SchemeFactory {
+ public add_partitions_pspec_argsStandardScheme getScheme() {
+ return new add_partitions_pspec_argsStandardScheme();
}
}
- private static class append_partition_argsStandardScheme extends StandardScheme<append_partition_args> {
+ private static class add_partitions_pspec_argsStandardScheme extends StandardScheme<add_partitions_pspec_args> {
- public void read(org.apache.thrift.protocol.TProtocol iprot, append_partition_args struct) throws org.apache.thrift.TException {
+ public void read(org.apache.thrift.protocol.TProtocol iprot, add_partitions_pspec_args struct) throws org.apache.thrift.TException {
org.apache.thrift.protocol.TField schemeField;
iprot.readStructBegin();
while (true)
@@ -40101,36 +40241,21 @@ public class ThriftHiveMetastore {
break;
}
switch (schemeField.id) {
- case 1: // DB_NAME
- if (schemeField.type == org.apache.thrift.protocol.TType.STRING) {
- struct.db_name = iprot.readString();
- struct.setDb_nameIsSet(true);
- } else {
- org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type);
- }
- break;
- case 2: // TBL_NAME
- if (schemeField.type == org.apache.thrift.protocol.TType.STRING) {
- struct.tbl_name = iprot.readString();
- struct.setTbl_nameIsSet(true);
- } else {
- org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type);
- }
- break;
- case 3: // PART_VALS
+ case 1: // NEW_PARTS
if (schemeField.type == org.apache.thrift.protocol.TType.LIST) {
{
- org.apache.thrift.protocol.TList _list548 = iprot.readListBegin();
- struct.part_vals = new ArrayList<String>(_list548.size);
- for (int _i549 = 0; _i549 < _list548.size; ++_i549)
+ org.apache.thrift.protocol.TList _list582 = iprot.readListBegin();
+ struct.new_parts = new ArrayList<PartitionSpec>(_list582.size);
+ for (int _i583 = 0; _i583 < _list582.size; ++_i583)
{
- String _elem550; // required
- _elem550 = iprot.readString();
- struct.part_vals.add(_elem550);
+ PartitionSpec _elem584; // required
+ _elem584 = new PartitionSpec();
+ _elem584.read(iprot);
+ struct.new_parts.add(_elem584);
}
iprot.readListEnd();
}
- struct.setPart_valsIsSet(true);
+ struct.setNew_partsIsSet(true);
} else {
org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type);
}
@@ -40144,27 +40269,17 @@ public class ThriftHiveMetastore {
struct.validate();
}
- public void write(org.apache.thrift.protocol.TProtocol oprot, append_partition_args struct) throws org.apache.thrift.TException {
+ public void write(org.apache.thrift.protocol.TProtocol oprot, add_partitions_pspec_args struct) throws org.apache.thrift.TException {
struct.validate();
oprot.writeStructBegin(STRUCT_DESC);
- if (struct.db_name != null) {
- oprot.writeFieldBegin(DB_NAME_FIELD_DESC);
- oprot.writeString(struct.db_name);
- oprot.writeFieldEnd();
- }
- if (struct.tbl_name != null) {
- oprot.writeFieldBegin(TBL_NAME_FIELD_DESC);
- oprot.writeString(struct.tbl_name);
- oprot.writeFieldEnd();
- }
- if (struct.part_vals != null) {
- oprot.writeFieldBegin(PART_VALS_FIELD_DESC);
+ if (struct.new_parts != null) {
+ oprot.writeFieldBegin(NEW_PARTS_FIELD_DESC);
{
- oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.part_vals.size()));
- for (String _iter551 : struct.part_vals)
+ oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.new_parts.size()));
+ for (PartitionSpec _iter585 : struct.new_parts)
{
- oprot.writeString(_iter551);
+ _iter585.write(oprot);
}
oprot.writeListEnd();
}
@@ -40176,90 +40291,71 @@ public class ThriftHiveMetastore {
}
- private static class append_partition_argsTupleSchemeFactory implements SchemeFactory {
- public append_partition_argsTupleScheme getScheme() {
- return new append_partition_argsTupleScheme();
+ private static class add_partitions_pspec_argsTupleSchemeFactory implements SchemeFactory {
+ public add_partitions_pspec_argsTupleScheme getScheme() {
+ return new add_partitions_pspec_argsTupleScheme();
}
}
- private static class append_partition_argsTupleScheme extends TupleScheme<append_partition_args> {
+ private static class add_partitions_pspec_argsTupleScheme extends TupleScheme<add_partitions_pspec_args> {
@Override
- public void write(org.apache.thrift.protocol.TProtocol prot, append_partition_args struct) throws org.apache.thrift.TException {
+ public void write(org.apache.thrift.protocol.TProtocol prot, add_partitions_pspec_args struct) throws org.apache.thrift.TException {
TTupleProtocol oprot = (TTupleProtocol) prot;
BitSet optionals = new BitSet();
- if (struct.isSetDb_name()) {
+ if (struct.isSetNew_parts()) {
optionals.set(0);
}
- if (struct.isSetTbl_name()) {
- optionals.set(1);
- }
- if (struct.isSetPart_vals()) {
- optionals.set(2);
- }
- oprot.writeBitSet(optionals, 3);
- if (struct.isSetDb_name()) {
- oprot.writeString(struct.db_name);
- }
- if (struct.isSetTbl_name()) {
- oprot.writeString(struct.tbl_name);
- }
- if (struct.isSetPart_vals()) {
+ oprot.writeBitSet(optionals, 1);
+ if (struct.isSetNew_parts()) {
{
- oprot.writeI32(struct.part_vals.size());
- for (String _iter552 : struct.part_vals)
+ oprot.writeI32(struct.new_parts.size());
+ for (PartitionSpec _iter586 : struct.new_parts)
{
- oprot.writeString(_iter552);
+ _iter586.write(oprot);
}
}
}
}
@Override
- public void read(org.apache.thrift.protocol.TProtocol prot, append_partition_args struct) throws org.apache.thrift.TException {
+ public void read(org.apache.thrift.protocol.TProtocol prot, add_partitions_pspec_args struct) throws org.apache.thrift.TException {
TTupleProtocol iprot = (TTupleProtocol) prot;
- BitSet incoming = iprot.readBitSet(3);
+ BitSet incoming = iprot.readBitSet(1);
if (incoming.get(0)) {
- struct.db_name = iprot.readString();
- struct.setDb_nameIsSet(true);
- }
- if (incoming.get(1)) {
- struct.tbl_name = iprot.readString();
- struct.setTbl_nameIsSet(true);
- }
- if (incoming.get(2)) {
{
- org.apache.thrift.protocol.TList _list553 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32());
- struct.part_vals = new ArrayList<String>(_list553.size);
- for (int _i554 = 0; _i554 < _list553.size; ++_i554)
+ org.apache.thrift.protocol.TList _list587 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32());
+ struct.new_parts = new ArrayList<PartitionSpec>(_list587.size);
+ for (int _i588 = 0; _i588 < _list587.size; ++_i588)
{
- String _elem555; // required
- _elem555 = iprot.readString();
- struct.part_vals.add(_elem555);
+ PartitionSpec _elem589; // required
+ _elem589 = new PartitionSpec();
+ _elem589.read(iprot);
+ struct.new_parts.add(_elem589);
}
}
- struct.setPart_valsIsSet(true);
+ struct.setNew_partsIsSet(true);
}
}
}
}
- public static class append_partition_result implements org.apache.thrift.TBase<append_partition_result, append_partition_result._Fields>, java.io.Serializable, Cloneable {
- private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("append_partition_result");
+ public static class add_partitions_pspec_result implements org.apache.thrift.TBase<add_partitions_pspec_result, add_partitions_pspec_result._Fields>, java.io.Serializable, Cloneable {
+ private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("add_partitions_pspec_result");
- private static final org.apache.thrift.protocol.TField SUCCESS_FIELD_DESC = new org.apache.thrift.protocol.TField("success", org.apache.thrift.protocol.TType.STRUCT, (short)0);
+ private static final org.apache.thrift.protocol.TField SUCCESS_FIELD_DESC = new org.apache.thrift.protocol.TField("success", org.apache.thrift.protocol.TType.I32, (short)0);
private static final org.apache.thrift.protocol.TField O1_FIELD_DESC = new org.apache.thrift.protocol.TField("o1", org.apache.thrift.protocol.TType.STRUCT, (short)1);
private static final org.apache.thrift.protocol.TField O2_FIELD_DESC = new org.apache.thrift.protocol.TField("o2", org.apache.thrift.protocol.TType.STRUCT, (short)2);
private static final org.apache.thrift.protocol.TField O3_FIELD_DESC = new org.apache.thrift.protocol.TField("o3", org.apache.thrift.protocol.TType.STRUCT, (short)3);
private static final Map<Class<? extends IScheme>, SchemeFactory> schemes = new HashMap<Class<? extends IScheme>, SchemeFactory>();
static {
- schemes.put(StandardScheme.class, new append_partition_resultStandardSchemeFactory());
- schemes.put(TupleScheme.class, new append_partition_resultTupleSchemeFactory());
+ schemes.put(StandardScheme.class, new add_partitions_pspec_resultStandardSchemeFactory());
+ schemes.put(TupleScheme.class, new add_partitions_pspec_resultTupleSchemeFactory());
}
- private Partition success; // required
+ private int success; // required
private InvalidObjectException o1; // required
private AlreadyExistsException o2; // required
private MetaException o3; // required
@@ -40332,11 +40428,13 @@ public class ThriftHiveMetastore {
}
// isset id assignments
+ private static final int __SUCCESS_ISSET_ID = 0;
+ private byte __isset_bitfield = 0;
public static final Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap;
static {
Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class);
tmpMap.put(_Fields.SUCCESS, new org.apache.thrift.meta_data.FieldMetaData("success", org.apache.thrift.TFieldRequirementType.DEFAULT,
- new org.apache.thrift.meta_data.StructMetaData(org.apache.thrift.protocol.TType.STRUCT, Partition.class)));
+ new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.I32)));
tmpMap.put(_Fields.O1, new org.apache.thrift.meta_data.FieldMetaData("o1", org.apache.thrift.TFieldRequirementType.DEFAULT,
[... 23007 lines stripped ...]