You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by am...@apache.org on 2011/09/05 10:47:06 UTC
svn commit: r1165207 [18/19] - in /hive/trunk: ./ cli/ eclipse-templates/
ivy/ jdbc/ lib/ metastore/ metastore/src/gen/thrift/gen-cpp/
metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/
metastore/src/gen/thrift/gen-php/hive_met...
Modified: hive/trunk/service/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/service/ThriftHive.java
URL: http://svn.apache.org/viewvc/hive/trunk/service/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/service/ThriftHive.java?rev=1165207&r1=1165206&r2=1165207&view=diff
==============================================================================
--- hive/trunk/service/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/service/ThriftHive.java (original)
+++ hive/trunk/service/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/service/ThriftHive.java Mon Sep 5 08:47:02 2011
@@ -1,5 +1,5 @@
/**
- * Autogenerated by Thrift
+ * Autogenerated by Thrift Compiler (0.7.0)
*
* DO NOT EDIT UNLESS YOU ARE SURE THAT YOU KNOW WHAT YOU ARE DOING
*/
@@ -20,1070 +20,787 @@ import java.util.Arrays;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
-import org.apache.thrift.*;
-import org.apache.thrift.async.*;
-import org.apache.thrift.meta_data.*;
-import org.apache.thrift.transport.*;
-import org.apache.thrift.protocol.*;
-
public class ThriftHive {
public interface Iface extends org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface {
- public void execute(String query) throws HiveServerException, TException;
+ public void execute(String query) throws HiveServerException, org.apache.thrift.TException;
- public String fetchOne() throws HiveServerException, TException;
+ public String fetchOne() throws HiveServerException, org.apache.thrift.TException;
- public List<String> fetchN(int numRows) throws HiveServerException, TException;
+ public List<String> fetchN(int numRows) throws HiveServerException, org.apache.thrift.TException;
- public List<String> fetchAll() throws HiveServerException, TException;
+ public List<String> fetchAll() throws HiveServerException, org.apache.thrift.TException;
- public org.apache.hadoop.hive.metastore.api.Schema getSchema() throws HiveServerException, TException;
+ public org.apache.hadoop.hive.metastore.api.Schema getSchema() throws HiveServerException, org.apache.thrift.TException;
- public org.apache.hadoop.hive.metastore.api.Schema getThriftSchema() throws HiveServerException, TException;
+ public org.apache.hadoop.hive.metastore.api.Schema getThriftSchema() throws HiveServerException, org.apache.thrift.TException;
- public HiveClusterStatus getClusterStatus() throws HiveServerException, TException;
+ public HiveClusterStatus getClusterStatus() throws HiveServerException, org.apache.thrift.TException;
- public org.apache.hadoop.hive.ql.plan.api.QueryPlan getQueryPlan() throws HiveServerException, TException;
+ public org.apache.hadoop.hive.ql.plan.api.QueryPlan getQueryPlan() throws HiveServerException, org.apache.thrift.TException;
- public void clean() throws TException;
+ public void clean() throws org.apache.thrift.TException;
}
public interface AsyncIface extends org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore .AsyncIface {
- public void execute(String query, AsyncMethodCallback<AsyncClient.execute_call> resultHandler) throws TException;
+ public void execute(String query, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.execute_call> resultHandler) throws org.apache.thrift.TException;
- public void fetchOne(AsyncMethodCallback<AsyncClient.fetchOne_call> resultHandler) throws TException;
+ public void fetchOne(org.apache.thrift.async.AsyncMethodCallback<AsyncClient.fetchOne_call> resultHandler) throws org.apache.thrift.TException;
- public void fetchN(int numRows, AsyncMethodCallback<AsyncClient.fetchN_call> resultHandler) throws TException;
+ public void fetchN(int numRows, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.fetchN_call> resultHandler) throws org.apache.thrift.TException;
- public void fetchAll(AsyncMethodCallback<AsyncClient.fetchAll_call> resultHandler) throws TException;
+ public void fetchAll(org.apache.thrift.async.AsyncMethodCallback<AsyncClient.fetchAll_call> resultHandler) throws org.apache.thrift.TException;
- public void getSchema(AsyncMethodCallback<AsyncClient.getSchema_call> resultHandler) throws TException;
+ public void getSchema(org.apache.thrift.async.AsyncMethodCallback<AsyncClient.getSchema_call> resultHandler) throws org.apache.thrift.TException;
- public void getThriftSchema(AsyncMethodCallback<AsyncClient.getThriftSchema_call> resultHandler) throws TException;
+ public void getThriftSchema(org.apache.thrift.async.AsyncMethodCallback<AsyncClient.getThriftSchema_call> resultHandler) throws org.apache.thrift.TException;
- public void getClusterStatus(AsyncMethodCallback<AsyncClient.getClusterStatus_call> resultHandler) throws TException;
+ public void getClusterStatus(org.apache.thrift.async.AsyncMethodCallback<AsyncClient.getClusterStatus_call> resultHandler) throws org.apache.thrift.TException;
- public void getQueryPlan(AsyncMethodCallback<AsyncClient.getQueryPlan_call> resultHandler) throws TException;
+ public void getQueryPlan(org.apache.thrift.async.AsyncMethodCallback<AsyncClient.getQueryPlan_call> resultHandler) throws org.apache.thrift.TException;
- public void clean(AsyncMethodCallback<AsyncClient.clean_call> resultHandler) throws TException;
+ public void clean(org.apache.thrift.async.AsyncMethodCallback<AsyncClient.clean_call> resultHandler) throws org.apache.thrift.TException;
}
- public static class Client extends org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Client implements TServiceClient, Iface {
- public static class Factory implements TServiceClientFactory<Client> {
+ public static class Client extends org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Client implements Iface {
+ public static class Factory implements org.apache.thrift.TServiceClientFactory<Client> {
public Factory() {}
- public Client getClient(TProtocol prot) {
+ public Client getClient(org.apache.thrift.protocol.TProtocol prot) {
return new Client(prot);
}
- public Client getClient(TProtocol iprot, TProtocol oprot) {
+ public Client getClient(org.apache.thrift.protocol.TProtocol iprot, org.apache.thrift.protocol.TProtocol oprot) {
return new Client(iprot, oprot);
}
}
- public Client(TProtocol prot)
+ public Client(org.apache.thrift.protocol.TProtocol prot)
{
- this(prot, prot);
+ super(prot, prot);
}
- public Client(TProtocol iprot, TProtocol oprot)
- {
+ public Client(org.apache.thrift.protocol.TProtocol iprot, org.apache.thrift.protocol.TProtocol oprot) {
super(iprot, oprot);
}
- public void execute(String query) throws HiveServerException, TException
+ public void execute(String query) throws HiveServerException, org.apache.thrift.TException
{
send_execute(query);
recv_execute();
}
- public void send_execute(String query) throws TException
+ public void send_execute(String query) throws org.apache.thrift.TException
{
- oprot_.writeMessageBegin(new TMessage("execute", TMessageType.CALL, ++seqid_));
execute_args args = new execute_args();
args.setQuery(query);
- args.write(oprot_);
- oprot_.writeMessageEnd();
- oprot_.getTransport().flush();
+ sendBase("execute", args);
}
- public void recv_execute() throws HiveServerException, TException
+ public void recv_execute() throws HiveServerException, org.apache.thrift.TException
{
- TMessage msg = iprot_.readMessageBegin();
- if (msg.type == TMessageType.EXCEPTION) {
- TApplicationException x = TApplicationException.read(iprot_);
- iprot_.readMessageEnd();
- throw x;
- }
- if (msg.seqid != seqid_) {
- throw new TApplicationException(TApplicationException.BAD_SEQUENCE_ID, "execute failed: out of sequence response");
- }
execute_result result = new execute_result();
- result.read(iprot_);
- iprot_.readMessageEnd();
+ receiveBase(result, "execute");
if (result.ex != null) {
throw result.ex;
}
return;
}
- public String fetchOne() throws HiveServerException, TException
+ public String fetchOne() throws HiveServerException, org.apache.thrift.TException
{
send_fetchOne();
return recv_fetchOne();
}
- public void send_fetchOne() throws TException
+ public void send_fetchOne() throws org.apache.thrift.TException
{
- oprot_.writeMessageBegin(new TMessage("fetchOne", TMessageType.CALL, ++seqid_));
fetchOne_args args = new fetchOne_args();
- args.write(oprot_);
- oprot_.writeMessageEnd();
- oprot_.getTransport().flush();
+ sendBase("fetchOne", args);
}
- public String recv_fetchOne() throws HiveServerException, TException
+ public String recv_fetchOne() throws HiveServerException, org.apache.thrift.TException
{
- TMessage msg = iprot_.readMessageBegin();
- if (msg.type == TMessageType.EXCEPTION) {
- TApplicationException x = TApplicationException.read(iprot_);
- iprot_.readMessageEnd();
- throw x;
- }
- if (msg.seqid != seqid_) {
- throw new TApplicationException(TApplicationException.BAD_SEQUENCE_ID, "fetchOne failed: out of sequence response");
- }
fetchOne_result result = new fetchOne_result();
- result.read(iprot_);
- iprot_.readMessageEnd();
+ receiveBase(result, "fetchOne");
if (result.isSetSuccess()) {
return result.success;
}
if (result.ex != null) {
throw result.ex;
}
- throw new TApplicationException(TApplicationException.MISSING_RESULT, "fetchOne failed: unknown result");
+ throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "fetchOne failed: unknown result");
}
- public List<String> fetchN(int numRows) throws HiveServerException, TException
+ public List<String> fetchN(int numRows) throws HiveServerException, org.apache.thrift.TException
{
send_fetchN(numRows);
return recv_fetchN();
}
- public void send_fetchN(int numRows) throws TException
+ public void send_fetchN(int numRows) throws org.apache.thrift.TException
{
- oprot_.writeMessageBegin(new TMessage("fetchN", TMessageType.CALL, ++seqid_));
fetchN_args args = new fetchN_args();
args.setNumRows(numRows);
- args.write(oprot_);
- oprot_.writeMessageEnd();
- oprot_.getTransport().flush();
+ sendBase("fetchN", args);
}
- public List<String> recv_fetchN() throws HiveServerException, TException
+ public List<String> recv_fetchN() throws HiveServerException, org.apache.thrift.TException
{
- TMessage msg = iprot_.readMessageBegin();
- if (msg.type == TMessageType.EXCEPTION) {
- TApplicationException x = TApplicationException.read(iprot_);
- iprot_.readMessageEnd();
- throw x;
- }
- if (msg.seqid != seqid_) {
- throw new TApplicationException(TApplicationException.BAD_SEQUENCE_ID, "fetchN failed: out of sequence response");
- }
fetchN_result result = new fetchN_result();
- result.read(iprot_);
- iprot_.readMessageEnd();
+ receiveBase(result, "fetchN");
if (result.isSetSuccess()) {
return result.success;
}
if (result.ex != null) {
throw result.ex;
}
- throw new TApplicationException(TApplicationException.MISSING_RESULT, "fetchN failed: unknown result");
+ throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "fetchN failed: unknown result");
}
- public List<String> fetchAll() throws HiveServerException, TException
+ public List<String> fetchAll() throws HiveServerException, org.apache.thrift.TException
{
send_fetchAll();
return recv_fetchAll();
}
- public void send_fetchAll() throws TException
+ public void send_fetchAll() throws org.apache.thrift.TException
{
- oprot_.writeMessageBegin(new TMessage("fetchAll", TMessageType.CALL, ++seqid_));
fetchAll_args args = new fetchAll_args();
- args.write(oprot_);
- oprot_.writeMessageEnd();
- oprot_.getTransport().flush();
+ sendBase("fetchAll", args);
}
- public List<String> recv_fetchAll() throws HiveServerException, TException
+ public List<String> recv_fetchAll() throws HiveServerException, org.apache.thrift.TException
{
- TMessage msg = iprot_.readMessageBegin();
- if (msg.type == TMessageType.EXCEPTION) {
- TApplicationException x = TApplicationException.read(iprot_);
- iprot_.readMessageEnd();
- throw x;
- }
- if (msg.seqid != seqid_) {
- throw new TApplicationException(TApplicationException.BAD_SEQUENCE_ID, "fetchAll failed: out of sequence response");
- }
fetchAll_result result = new fetchAll_result();
- result.read(iprot_);
- iprot_.readMessageEnd();
+ receiveBase(result, "fetchAll");
if (result.isSetSuccess()) {
return result.success;
}
if (result.ex != null) {
throw result.ex;
}
- throw new TApplicationException(TApplicationException.MISSING_RESULT, "fetchAll failed: unknown result");
+ throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "fetchAll failed: unknown result");
}
- public org.apache.hadoop.hive.metastore.api.Schema getSchema() throws HiveServerException, TException
+ public org.apache.hadoop.hive.metastore.api.Schema getSchema() throws HiveServerException, org.apache.thrift.TException
{
send_getSchema();
return recv_getSchema();
}
- public void send_getSchema() throws TException
+ public void send_getSchema() throws org.apache.thrift.TException
{
- oprot_.writeMessageBegin(new TMessage("getSchema", TMessageType.CALL, ++seqid_));
getSchema_args args = new getSchema_args();
- args.write(oprot_);
- oprot_.writeMessageEnd();
- oprot_.getTransport().flush();
+ sendBase("getSchema", args);
}
- public org.apache.hadoop.hive.metastore.api.Schema recv_getSchema() throws HiveServerException, TException
+ public org.apache.hadoop.hive.metastore.api.Schema recv_getSchema() throws HiveServerException, org.apache.thrift.TException
{
- TMessage msg = iprot_.readMessageBegin();
- if (msg.type == TMessageType.EXCEPTION) {
- TApplicationException x = TApplicationException.read(iprot_);
- iprot_.readMessageEnd();
- throw x;
- }
- if (msg.seqid != seqid_) {
- throw new TApplicationException(TApplicationException.BAD_SEQUENCE_ID, "getSchema failed: out of sequence response");
- }
getSchema_result result = new getSchema_result();
- result.read(iprot_);
- iprot_.readMessageEnd();
+ receiveBase(result, "getSchema");
if (result.isSetSuccess()) {
return result.success;
}
if (result.ex != null) {
throw result.ex;
}
- throw new TApplicationException(TApplicationException.MISSING_RESULT, "getSchema failed: unknown result");
+ throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "getSchema failed: unknown result");
}
- public org.apache.hadoop.hive.metastore.api.Schema getThriftSchema() throws HiveServerException, TException
+ public org.apache.hadoop.hive.metastore.api.Schema getThriftSchema() throws HiveServerException, org.apache.thrift.TException
{
send_getThriftSchema();
return recv_getThriftSchema();
}
- public void send_getThriftSchema() throws TException
+ public void send_getThriftSchema() throws org.apache.thrift.TException
{
- oprot_.writeMessageBegin(new TMessage("getThriftSchema", TMessageType.CALL, ++seqid_));
getThriftSchema_args args = new getThriftSchema_args();
- args.write(oprot_);
- oprot_.writeMessageEnd();
- oprot_.getTransport().flush();
+ sendBase("getThriftSchema", args);
}
- public org.apache.hadoop.hive.metastore.api.Schema recv_getThriftSchema() throws HiveServerException, TException
+ public org.apache.hadoop.hive.metastore.api.Schema recv_getThriftSchema() throws HiveServerException, org.apache.thrift.TException
{
- TMessage msg = iprot_.readMessageBegin();
- if (msg.type == TMessageType.EXCEPTION) {
- TApplicationException x = TApplicationException.read(iprot_);
- iprot_.readMessageEnd();
- throw x;
- }
- if (msg.seqid != seqid_) {
- throw new TApplicationException(TApplicationException.BAD_SEQUENCE_ID, "getThriftSchema failed: out of sequence response");
- }
getThriftSchema_result result = new getThriftSchema_result();
- result.read(iprot_);
- iprot_.readMessageEnd();
+ receiveBase(result, "getThriftSchema");
if (result.isSetSuccess()) {
return result.success;
}
if (result.ex != null) {
throw result.ex;
}
- throw new TApplicationException(TApplicationException.MISSING_RESULT, "getThriftSchema failed: unknown result");
+ throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "getThriftSchema failed: unknown result");
}
- public HiveClusterStatus getClusterStatus() throws HiveServerException, TException
+ public HiveClusterStatus getClusterStatus() throws HiveServerException, org.apache.thrift.TException
{
send_getClusterStatus();
return recv_getClusterStatus();
}
- public void send_getClusterStatus() throws TException
+ public void send_getClusterStatus() throws org.apache.thrift.TException
{
- oprot_.writeMessageBegin(new TMessage("getClusterStatus", TMessageType.CALL, ++seqid_));
getClusterStatus_args args = new getClusterStatus_args();
- args.write(oprot_);
- oprot_.writeMessageEnd();
- oprot_.getTransport().flush();
+ sendBase("getClusterStatus", args);
}
- public HiveClusterStatus recv_getClusterStatus() throws HiveServerException, TException
+ public HiveClusterStatus recv_getClusterStatus() throws HiveServerException, org.apache.thrift.TException
{
- TMessage msg = iprot_.readMessageBegin();
- if (msg.type == TMessageType.EXCEPTION) {
- TApplicationException x = TApplicationException.read(iprot_);
- iprot_.readMessageEnd();
- throw x;
- }
- if (msg.seqid != seqid_) {
- throw new TApplicationException(TApplicationException.BAD_SEQUENCE_ID, "getClusterStatus failed: out of sequence response");
- }
getClusterStatus_result result = new getClusterStatus_result();
- result.read(iprot_);
- iprot_.readMessageEnd();
+ receiveBase(result, "getClusterStatus");
if (result.isSetSuccess()) {
return result.success;
}
if (result.ex != null) {
throw result.ex;
}
- throw new TApplicationException(TApplicationException.MISSING_RESULT, "getClusterStatus failed: unknown result");
+ throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "getClusterStatus failed: unknown result");
}
- public org.apache.hadoop.hive.ql.plan.api.QueryPlan getQueryPlan() throws HiveServerException, TException
+ public org.apache.hadoop.hive.ql.plan.api.QueryPlan getQueryPlan() throws HiveServerException, org.apache.thrift.TException
{
send_getQueryPlan();
return recv_getQueryPlan();
}
- public void send_getQueryPlan() throws TException
+ public void send_getQueryPlan() throws org.apache.thrift.TException
{
- oprot_.writeMessageBegin(new TMessage("getQueryPlan", TMessageType.CALL, ++seqid_));
getQueryPlan_args args = new getQueryPlan_args();
- args.write(oprot_);
- oprot_.writeMessageEnd();
- oprot_.getTransport().flush();
+ sendBase("getQueryPlan", args);
}
- public org.apache.hadoop.hive.ql.plan.api.QueryPlan recv_getQueryPlan() throws HiveServerException, TException
+ public org.apache.hadoop.hive.ql.plan.api.QueryPlan recv_getQueryPlan() throws HiveServerException, org.apache.thrift.TException
{
- TMessage msg = iprot_.readMessageBegin();
- if (msg.type == TMessageType.EXCEPTION) {
- TApplicationException x = TApplicationException.read(iprot_);
- iprot_.readMessageEnd();
- throw x;
- }
- if (msg.seqid != seqid_) {
- throw new TApplicationException(TApplicationException.BAD_SEQUENCE_ID, "getQueryPlan failed: out of sequence response");
- }
getQueryPlan_result result = new getQueryPlan_result();
- result.read(iprot_);
- iprot_.readMessageEnd();
+ receiveBase(result, "getQueryPlan");
if (result.isSetSuccess()) {
return result.success;
}
if (result.ex != null) {
throw result.ex;
}
- throw new TApplicationException(TApplicationException.MISSING_RESULT, "getQueryPlan failed: unknown result");
+ throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "getQueryPlan failed: unknown result");
}
- public void clean() throws TException
+ public void clean() throws org.apache.thrift.TException
{
send_clean();
recv_clean();
}
- public void send_clean() throws TException
+ public void send_clean() throws org.apache.thrift.TException
{
- oprot_.writeMessageBegin(new TMessage("clean", TMessageType.CALL, ++seqid_));
clean_args args = new clean_args();
- args.write(oprot_);
- oprot_.writeMessageEnd();
- oprot_.getTransport().flush();
+ sendBase("clean", args);
}
- public void recv_clean() throws TException
+ public void recv_clean() throws org.apache.thrift.TException
{
- TMessage msg = iprot_.readMessageBegin();
- if (msg.type == TMessageType.EXCEPTION) {
- TApplicationException x = TApplicationException.read(iprot_);
- iprot_.readMessageEnd();
- throw x;
- }
- if (msg.seqid != seqid_) {
- throw new TApplicationException(TApplicationException.BAD_SEQUENCE_ID, "clean failed: out of sequence response");
- }
clean_result result = new clean_result();
- result.read(iprot_);
- iprot_.readMessageEnd();
+ receiveBase(result, "clean");
return;
}
}
public static class AsyncClient extends org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.AsyncClient implements AsyncIface {
- public static class Factory implements TAsyncClientFactory<AsyncClient> {
- private TAsyncClientManager clientManager;
- private TProtocolFactory protocolFactory;
- public Factory(TAsyncClientManager clientManager, TProtocolFactory protocolFactory) {
+ public static class Factory implements org.apache.thrift.async.TAsyncClientFactory<AsyncClient> {
+ private org.apache.thrift.async.TAsyncClientManager clientManager;
+ private org.apache.thrift.protocol.TProtocolFactory protocolFactory;
+ public Factory(org.apache.thrift.async.TAsyncClientManager clientManager, org.apache.thrift.protocol.TProtocolFactory protocolFactory) {
this.clientManager = clientManager;
this.protocolFactory = protocolFactory;
}
- public AsyncClient getAsyncClient(TNonblockingTransport transport) {
+ public AsyncClient getAsyncClient(org.apache.thrift.transport.TNonblockingTransport transport) {
return new AsyncClient(protocolFactory, clientManager, transport);
}
}
- public AsyncClient(TProtocolFactory protocolFactory, TAsyncClientManager clientManager, TNonblockingTransport transport) {
+ public AsyncClient(org.apache.thrift.protocol.TProtocolFactory protocolFactory, org.apache.thrift.async.TAsyncClientManager clientManager, org.apache.thrift.transport.TNonblockingTransport transport) {
super(protocolFactory, clientManager, transport);
}
- public void execute(String query, AsyncMethodCallback<execute_call> resultHandler) throws TException {
+ public void execute(String query, org.apache.thrift.async.AsyncMethodCallback<execute_call> resultHandler) throws org.apache.thrift.TException {
checkReady();
- execute_call method_call = new execute_call(query, resultHandler, this, protocolFactory, transport);
- manager.call(method_call);
+ execute_call method_call = new execute_call(query, resultHandler, this, ___protocolFactory, ___transport);
+ this.___currentMethod = method_call;
+ ___manager.call(method_call);
}
- public static class execute_call extends TAsyncMethodCall {
+ public static class execute_call extends org.apache.thrift.async.TAsyncMethodCall {
private String query;
- public execute_call(String query, AsyncMethodCallback<execute_call> resultHandler, TAsyncClient client, TProtocolFactory protocolFactory, TNonblockingTransport transport) throws TException {
+ public execute_call(String query, org.apache.thrift.async.AsyncMethodCallback<execute_call> resultHandler, org.apache.thrift.async.TAsyncClient client, org.apache.thrift.protocol.TProtocolFactory protocolFactory, org.apache.thrift.transport.TNonblockingTransport transport) throws org.apache.thrift.TException {
super(client, protocolFactory, transport, resultHandler, false);
this.query = query;
}
- public void write_args(TProtocol prot) throws TException {
- prot.writeMessageBegin(new TMessage("execute", TMessageType.CALL, 0));
+ public void write_args(org.apache.thrift.protocol.TProtocol prot) throws org.apache.thrift.TException {
+ prot.writeMessageBegin(new org.apache.thrift.protocol.TMessage("execute", org.apache.thrift.protocol.TMessageType.CALL, 0));
execute_args args = new execute_args();
args.setQuery(query);
args.write(prot);
prot.writeMessageEnd();
}
- public void getResult() throws HiveServerException, TException {
- if (getState() != State.RESPONSE_READ) {
+ public void getResult() throws HiveServerException, org.apache.thrift.TException {
+ if (getState() != org.apache.thrift.async.TAsyncMethodCall.State.RESPONSE_READ) {
throw new IllegalStateException("Method call not finished!");
}
- TMemoryInputTransport memoryTransport = new TMemoryInputTransport(getFrameBuffer().array());
- TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport);
+ org.apache.thrift.transport.TMemoryInputTransport memoryTransport = new org.apache.thrift.transport.TMemoryInputTransport(getFrameBuffer().array());
+ org.apache.thrift.protocol.TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport);
(new Client(prot)).recv_execute();
}
}
- public void fetchOne(AsyncMethodCallback<fetchOne_call> resultHandler) throws TException {
+ public void fetchOne(org.apache.thrift.async.AsyncMethodCallback<fetchOne_call> resultHandler) throws org.apache.thrift.TException {
checkReady();
- fetchOne_call method_call = new fetchOne_call(resultHandler, this, protocolFactory, transport);
- manager.call(method_call);
+ fetchOne_call method_call = new fetchOne_call(resultHandler, this, ___protocolFactory, ___transport);
+ this.___currentMethod = method_call;
+ ___manager.call(method_call);
}
- public static class fetchOne_call extends TAsyncMethodCall {
- public fetchOne_call(AsyncMethodCallback<fetchOne_call> resultHandler, TAsyncClient client, TProtocolFactory protocolFactory, TNonblockingTransport transport) throws TException {
+ public static class fetchOne_call extends org.apache.thrift.async.TAsyncMethodCall {
+ public fetchOne_call(org.apache.thrift.async.AsyncMethodCallback<fetchOne_call> resultHandler, org.apache.thrift.async.TAsyncClient client, org.apache.thrift.protocol.TProtocolFactory protocolFactory, org.apache.thrift.transport.TNonblockingTransport transport) throws org.apache.thrift.TException {
super(client, protocolFactory, transport, resultHandler, false);
}
- public void write_args(TProtocol prot) throws TException {
- prot.writeMessageBegin(new TMessage("fetchOne", TMessageType.CALL, 0));
+ public void write_args(org.apache.thrift.protocol.TProtocol prot) throws org.apache.thrift.TException {
+ prot.writeMessageBegin(new org.apache.thrift.protocol.TMessage("fetchOne", org.apache.thrift.protocol.TMessageType.CALL, 0));
fetchOne_args args = new fetchOne_args();
args.write(prot);
prot.writeMessageEnd();
}
- public String getResult() throws HiveServerException, TException {
- if (getState() != State.RESPONSE_READ) {
+ public String getResult() throws HiveServerException, org.apache.thrift.TException {
+ if (getState() != org.apache.thrift.async.TAsyncMethodCall.State.RESPONSE_READ) {
throw new IllegalStateException("Method call not finished!");
}
- TMemoryInputTransport memoryTransport = new TMemoryInputTransport(getFrameBuffer().array());
- TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport);
+ org.apache.thrift.transport.TMemoryInputTransport memoryTransport = new org.apache.thrift.transport.TMemoryInputTransport(getFrameBuffer().array());
+ org.apache.thrift.protocol.TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport);
return (new Client(prot)).recv_fetchOne();
}
}
- public void fetchN(int numRows, AsyncMethodCallback<fetchN_call> resultHandler) throws TException {
+ public void fetchN(int numRows, org.apache.thrift.async.AsyncMethodCallback<fetchN_call> resultHandler) throws org.apache.thrift.TException {
checkReady();
- fetchN_call method_call = new fetchN_call(numRows, resultHandler, this, protocolFactory, transport);
- manager.call(method_call);
+ fetchN_call method_call = new fetchN_call(numRows, resultHandler, this, ___protocolFactory, ___transport);
+ this.___currentMethod = method_call;
+ ___manager.call(method_call);
}
- public static class fetchN_call extends TAsyncMethodCall {
+ public static class fetchN_call extends org.apache.thrift.async.TAsyncMethodCall {
private int numRows;
- public fetchN_call(int numRows, AsyncMethodCallback<fetchN_call> resultHandler, TAsyncClient client, TProtocolFactory protocolFactory, TNonblockingTransport transport) throws TException {
+ public fetchN_call(int numRows, org.apache.thrift.async.AsyncMethodCallback<fetchN_call> resultHandler, org.apache.thrift.async.TAsyncClient client, org.apache.thrift.protocol.TProtocolFactory protocolFactory, org.apache.thrift.transport.TNonblockingTransport transport) throws org.apache.thrift.TException {
super(client, protocolFactory, transport, resultHandler, false);
this.numRows = numRows;
}
- public void write_args(TProtocol prot) throws TException {
- prot.writeMessageBegin(new TMessage("fetchN", TMessageType.CALL, 0));
+ public void write_args(org.apache.thrift.protocol.TProtocol prot) throws org.apache.thrift.TException {
+ prot.writeMessageBegin(new org.apache.thrift.protocol.TMessage("fetchN", org.apache.thrift.protocol.TMessageType.CALL, 0));
fetchN_args args = new fetchN_args();
args.setNumRows(numRows);
args.write(prot);
prot.writeMessageEnd();
}
- public List<String> getResult() throws HiveServerException, TException {
- if (getState() != State.RESPONSE_READ) {
+ public List<String> getResult() throws HiveServerException, org.apache.thrift.TException {
+ if (getState() != org.apache.thrift.async.TAsyncMethodCall.State.RESPONSE_READ) {
throw new IllegalStateException("Method call not finished!");
}
- TMemoryInputTransport memoryTransport = new TMemoryInputTransport(getFrameBuffer().array());
- TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport);
+ org.apache.thrift.transport.TMemoryInputTransport memoryTransport = new org.apache.thrift.transport.TMemoryInputTransport(getFrameBuffer().array());
+ org.apache.thrift.protocol.TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport);
return (new Client(prot)).recv_fetchN();
}
}
- public void fetchAll(AsyncMethodCallback<fetchAll_call> resultHandler) throws TException {
+ public void fetchAll(org.apache.thrift.async.AsyncMethodCallback<fetchAll_call> resultHandler) throws org.apache.thrift.TException {
checkReady();
- fetchAll_call method_call = new fetchAll_call(resultHandler, this, protocolFactory, transport);
- manager.call(method_call);
+ fetchAll_call method_call = new fetchAll_call(resultHandler, this, ___protocolFactory, ___transport);
+ this.___currentMethod = method_call;
+ ___manager.call(method_call);
}
- public static class fetchAll_call extends TAsyncMethodCall {
- public fetchAll_call(AsyncMethodCallback<fetchAll_call> resultHandler, TAsyncClient client, TProtocolFactory protocolFactory, TNonblockingTransport transport) throws TException {
+ public static class fetchAll_call extends org.apache.thrift.async.TAsyncMethodCall {
+ public fetchAll_call(org.apache.thrift.async.AsyncMethodCallback<fetchAll_call> resultHandler, org.apache.thrift.async.TAsyncClient client, org.apache.thrift.protocol.TProtocolFactory protocolFactory, org.apache.thrift.transport.TNonblockingTransport transport) throws org.apache.thrift.TException {
super(client, protocolFactory, transport, resultHandler, false);
}
- public void write_args(TProtocol prot) throws TException {
- prot.writeMessageBegin(new TMessage("fetchAll", TMessageType.CALL, 0));
+ public void write_args(org.apache.thrift.protocol.TProtocol prot) throws org.apache.thrift.TException {
+ prot.writeMessageBegin(new org.apache.thrift.protocol.TMessage("fetchAll", org.apache.thrift.protocol.TMessageType.CALL, 0));
fetchAll_args args = new fetchAll_args();
args.write(prot);
prot.writeMessageEnd();
}
- public List<String> getResult() throws HiveServerException, TException {
- if (getState() != State.RESPONSE_READ) {
+ public List<String> getResult() throws HiveServerException, org.apache.thrift.TException {
+ if (getState() != org.apache.thrift.async.TAsyncMethodCall.State.RESPONSE_READ) {
throw new IllegalStateException("Method call not finished!");
}
- TMemoryInputTransport memoryTransport = new TMemoryInputTransport(getFrameBuffer().array());
- TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport);
+ org.apache.thrift.transport.TMemoryInputTransport memoryTransport = new org.apache.thrift.transport.TMemoryInputTransport(getFrameBuffer().array());
+ org.apache.thrift.protocol.TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport);
return (new Client(prot)).recv_fetchAll();
}
}
- public void getSchema(AsyncMethodCallback<getSchema_call> resultHandler) throws TException {
+ public void getSchema(org.apache.thrift.async.AsyncMethodCallback<getSchema_call> resultHandler) throws org.apache.thrift.TException {
checkReady();
- getSchema_call method_call = new getSchema_call(resultHandler, this, protocolFactory, transport);
- manager.call(method_call);
+ getSchema_call method_call = new getSchema_call(resultHandler, this, ___protocolFactory, ___transport);
+ this.___currentMethod = method_call;
+ ___manager.call(method_call);
}
- public static class getSchema_call extends TAsyncMethodCall {
- public getSchema_call(AsyncMethodCallback<getSchema_call> resultHandler, TAsyncClient client, TProtocolFactory protocolFactory, TNonblockingTransport transport) throws TException {
+ public static class getSchema_call extends org.apache.thrift.async.TAsyncMethodCall {
+ public getSchema_call(org.apache.thrift.async.AsyncMethodCallback<getSchema_call> resultHandler, org.apache.thrift.async.TAsyncClient client, org.apache.thrift.protocol.TProtocolFactory protocolFactory, org.apache.thrift.transport.TNonblockingTransport transport) throws org.apache.thrift.TException {
super(client, protocolFactory, transport, resultHandler, false);
}
- public void write_args(TProtocol prot) throws TException {
- prot.writeMessageBegin(new TMessage("getSchema", TMessageType.CALL, 0));
+ public void write_args(org.apache.thrift.protocol.TProtocol prot) throws org.apache.thrift.TException {
+ prot.writeMessageBegin(new org.apache.thrift.protocol.TMessage("getSchema", org.apache.thrift.protocol.TMessageType.CALL, 0));
getSchema_args args = new getSchema_args();
args.write(prot);
prot.writeMessageEnd();
}
- public org.apache.hadoop.hive.metastore.api.Schema getResult() throws HiveServerException, TException {
- if (getState() != State.RESPONSE_READ) {
+ public org.apache.hadoop.hive.metastore.api.Schema getResult() throws HiveServerException, org.apache.thrift.TException {
+ if (getState() != org.apache.thrift.async.TAsyncMethodCall.State.RESPONSE_READ) {
throw new IllegalStateException("Method call not finished!");
}
- TMemoryInputTransport memoryTransport = new TMemoryInputTransport(getFrameBuffer().array());
- TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport);
+ org.apache.thrift.transport.TMemoryInputTransport memoryTransport = new org.apache.thrift.transport.TMemoryInputTransport(getFrameBuffer().array());
+ org.apache.thrift.protocol.TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport);
return (new Client(prot)).recv_getSchema();
}
}
- public void getThriftSchema(AsyncMethodCallback<getThriftSchema_call> resultHandler) throws TException {
+ public void getThriftSchema(org.apache.thrift.async.AsyncMethodCallback<getThriftSchema_call> resultHandler) throws org.apache.thrift.TException {
checkReady();
- getThriftSchema_call method_call = new getThriftSchema_call(resultHandler, this, protocolFactory, transport);
- manager.call(method_call);
+ getThriftSchema_call method_call = new getThriftSchema_call(resultHandler, this, ___protocolFactory, ___transport);
+ this.___currentMethod = method_call;
+ ___manager.call(method_call);
}
- public static class getThriftSchema_call extends TAsyncMethodCall {
- public getThriftSchema_call(AsyncMethodCallback<getThriftSchema_call> resultHandler, TAsyncClient client, TProtocolFactory protocolFactory, TNonblockingTransport transport) throws TException {
+ public static class getThriftSchema_call extends org.apache.thrift.async.TAsyncMethodCall {
+ public getThriftSchema_call(org.apache.thrift.async.AsyncMethodCallback<getThriftSchema_call> resultHandler, org.apache.thrift.async.TAsyncClient client, org.apache.thrift.protocol.TProtocolFactory protocolFactory, org.apache.thrift.transport.TNonblockingTransport transport) throws org.apache.thrift.TException {
super(client, protocolFactory, transport, resultHandler, false);
}
- public void write_args(TProtocol prot) throws TException {
- prot.writeMessageBegin(new TMessage("getThriftSchema", TMessageType.CALL, 0));
+ public void write_args(org.apache.thrift.protocol.TProtocol prot) throws org.apache.thrift.TException {
+ prot.writeMessageBegin(new org.apache.thrift.protocol.TMessage("getThriftSchema", org.apache.thrift.protocol.TMessageType.CALL, 0));
getThriftSchema_args args = new getThriftSchema_args();
args.write(prot);
prot.writeMessageEnd();
}
- public org.apache.hadoop.hive.metastore.api.Schema getResult() throws HiveServerException, TException {
- if (getState() != State.RESPONSE_READ) {
+ public org.apache.hadoop.hive.metastore.api.Schema getResult() throws HiveServerException, org.apache.thrift.TException {
+ if (getState() != org.apache.thrift.async.TAsyncMethodCall.State.RESPONSE_READ) {
throw new IllegalStateException("Method call not finished!");
}
- TMemoryInputTransport memoryTransport = new TMemoryInputTransport(getFrameBuffer().array());
- TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport);
+ org.apache.thrift.transport.TMemoryInputTransport memoryTransport = new org.apache.thrift.transport.TMemoryInputTransport(getFrameBuffer().array());
+ org.apache.thrift.protocol.TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport);
return (new Client(prot)).recv_getThriftSchema();
}
}
- public void getClusterStatus(AsyncMethodCallback<getClusterStatus_call> resultHandler) throws TException {
+ public void getClusterStatus(org.apache.thrift.async.AsyncMethodCallback<getClusterStatus_call> resultHandler) throws org.apache.thrift.TException {
checkReady();
- getClusterStatus_call method_call = new getClusterStatus_call(resultHandler, this, protocolFactory, transport);
- manager.call(method_call);
+ getClusterStatus_call method_call = new getClusterStatus_call(resultHandler, this, ___protocolFactory, ___transport);
+ this.___currentMethod = method_call;
+ ___manager.call(method_call);
}
- public static class getClusterStatus_call extends TAsyncMethodCall {
- public getClusterStatus_call(AsyncMethodCallback<getClusterStatus_call> resultHandler, TAsyncClient client, TProtocolFactory protocolFactory, TNonblockingTransport transport) throws TException {
+ public static class getClusterStatus_call extends org.apache.thrift.async.TAsyncMethodCall {
+ public getClusterStatus_call(org.apache.thrift.async.AsyncMethodCallback<getClusterStatus_call> resultHandler, org.apache.thrift.async.TAsyncClient client, org.apache.thrift.protocol.TProtocolFactory protocolFactory, org.apache.thrift.transport.TNonblockingTransport transport) throws org.apache.thrift.TException {
super(client, protocolFactory, transport, resultHandler, false);
}
- public void write_args(TProtocol prot) throws TException {
- prot.writeMessageBegin(new TMessage("getClusterStatus", TMessageType.CALL, 0));
+ public void write_args(org.apache.thrift.protocol.TProtocol prot) throws org.apache.thrift.TException {
+ prot.writeMessageBegin(new org.apache.thrift.protocol.TMessage("getClusterStatus", org.apache.thrift.protocol.TMessageType.CALL, 0));
getClusterStatus_args args = new getClusterStatus_args();
args.write(prot);
prot.writeMessageEnd();
}
- public HiveClusterStatus getResult() throws HiveServerException, TException {
- if (getState() != State.RESPONSE_READ) {
+ public HiveClusterStatus getResult() throws HiveServerException, org.apache.thrift.TException {
+ if (getState() != org.apache.thrift.async.TAsyncMethodCall.State.RESPONSE_READ) {
throw new IllegalStateException("Method call not finished!");
}
- TMemoryInputTransport memoryTransport = new TMemoryInputTransport(getFrameBuffer().array());
- TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport);
+ org.apache.thrift.transport.TMemoryInputTransport memoryTransport = new org.apache.thrift.transport.TMemoryInputTransport(getFrameBuffer().array());
+ org.apache.thrift.protocol.TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport);
return (new Client(prot)).recv_getClusterStatus();
}
}
- public void getQueryPlan(AsyncMethodCallback<getQueryPlan_call> resultHandler) throws TException {
+ public void getQueryPlan(org.apache.thrift.async.AsyncMethodCallback<getQueryPlan_call> resultHandler) throws org.apache.thrift.TException {
checkReady();
- getQueryPlan_call method_call = new getQueryPlan_call(resultHandler, this, protocolFactory, transport);
- manager.call(method_call);
+ getQueryPlan_call method_call = new getQueryPlan_call(resultHandler, this, ___protocolFactory, ___transport);
+ this.___currentMethod = method_call;
+ ___manager.call(method_call);
}
- public static class getQueryPlan_call extends TAsyncMethodCall {
- public getQueryPlan_call(AsyncMethodCallback<getQueryPlan_call> resultHandler, TAsyncClient client, TProtocolFactory protocolFactory, TNonblockingTransport transport) throws TException {
+ public static class getQueryPlan_call extends org.apache.thrift.async.TAsyncMethodCall {
+ public getQueryPlan_call(org.apache.thrift.async.AsyncMethodCallback<getQueryPlan_call> resultHandler, org.apache.thrift.async.TAsyncClient client, org.apache.thrift.protocol.TProtocolFactory protocolFactory, org.apache.thrift.transport.TNonblockingTransport transport) throws org.apache.thrift.TException {
super(client, protocolFactory, transport, resultHandler, false);
}
- public void write_args(TProtocol prot) throws TException {
- prot.writeMessageBegin(new TMessage("getQueryPlan", TMessageType.CALL, 0));
+ public void write_args(org.apache.thrift.protocol.TProtocol prot) throws org.apache.thrift.TException {
+ prot.writeMessageBegin(new org.apache.thrift.protocol.TMessage("getQueryPlan", org.apache.thrift.protocol.TMessageType.CALL, 0));
getQueryPlan_args args = new getQueryPlan_args();
args.write(prot);
prot.writeMessageEnd();
}
- public org.apache.hadoop.hive.ql.plan.api.QueryPlan getResult() throws HiveServerException, TException {
- if (getState() != State.RESPONSE_READ) {
+ public org.apache.hadoop.hive.ql.plan.api.QueryPlan getResult() throws HiveServerException, org.apache.thrift.TException {
+ if (getState() != org.apache.thrift.async.TAsyncMethodCall.State.RESPONSE_READ) {
throw new IllegalStateException("Method call not finished!");
}
- TMemoryInputTransport memoryTransport = new TMemoryInputTransport(getFrameBuffer().array());
- TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport);
+ org.apache.thrift.transport.TMemoryInputTransport memoryTransport = new org.apache.thrift.transport.TMemoryInputTransport(getFrameBuffer().array());
+ org.apache.thrift.protocol.TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport);
return (new Client(prot)).recv_getQueryPlan();
}
}
- public void clean(AsyncMethodCallback<clean_call> resultHandler) throws TException {
+ public void clean(org.apache.thrift.async.AsyncMethodCallback<clean_call> resultHandler) throws org.apache.thrift.TException {
checkReady();
- clean_call method_call = new clean_call(resultHandler, this, protocolFactory, transport);
- manager.call(method_call);
+ clean_call method_call = new clean_call(resultHandler, this, ___protocolFactory, ___transport);
+ this.___currentMethod = method_call;
+ ___manager.call(method_call);
}
- public static class clean_call extends TAsyncMethodCall {
- public clean_call(AsyncMethodCallback<clean_call> resultHandler, TAsyncClient client, TProtocolFactory protocolFactory, TNonblockingTransport transport) throws TException {
+ public static class clean_call extends org.apache.thrift.async.TAsyncMethodCall {
+ public clean_call(org.apache.thrift.async.AsyncMethodCallback<clean_call> resultHandler, org.apache.thrift.async.TAsyncClient client, org.apache.thrift.protocol.TProtocolFactory protocolFactory, org.apache.thrift.transport.TNonblockingTransport transport) throws org.apache.thrift.TException {
super(client, protocolFactory, transport, resultHandler, false);
}
- public void write_args(TProtocol prot) throws TException {
- prot.writeMessageBegin(new TMessage("clean", TMessageType.CALL, 0));
+ public void write_args(org.apache.thrift.protocol.TProtocol prot) throws org.apache.thrift.TException {
+ prot.writeMessageBegin(new org.apache.thrift.protocol.TMessage("clean", org.apache.thrift.protocol.TMessageType.CALL, 0));
clean_args args = new clean_args();
args.write(prot);
prot.writeMessageEnd();
}
- public void getResult() throws TException {
- if (getState() != State.RESPONSE_READ) {
+ public void getResult() throws org.apache.thrift.TException {
+ if (getState() != org.apache.thrift.async.TAsyncMethodCall.State.RESPONSE_READ) {
throw new IllegalStateException("Method call not finished!");
}
- TMemoryInputTransport memoryTransport = new TMemoryInputTransport(getFrameBuffer().array());
- TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport);
+ org.apache.thrift.transport.TMemoryInputTransport memoryTransport = new org.apache.thrift.transport.TMemoryInputTransport(getFrameBuffer().array());
+ org.apache.thrift.protocol.TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport);
(new Client(prot)).recv_clean();
}
}
}
- public static class Processor extends org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Processor implements TProcessor {
+ public static class Processor<I extends Iface> extends org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Processor implements org.apache.thrift.TProcessor {
private static final Logger LOGGER = LoggerFactory.getLogger(Processor.class.getName());
- public Processor(Iface iface)
- {
- super(iface);
- iface_ = iface;
- processMap_.put("execute", new execute());
- processMap_.put("fetchOne", new fetchOne());
- processMap_.put("fetchN", new fetchN());
- processMap_.put("fetchAll", new fetchAll());
- processMap_.put("getSchema", new getSchema());
- processMap_.put("getThriftSchema", new getThriftSchema());
- processMap_.put("getClusterStatus", new getClusterStatus());
- processMap_.put("getQueryPlan", new getQueryPlan());
- processMap_.put("clean", new clean());
+ public Processor(I iface) {
+ super(iface, getProcessMap(new HashMap<String, org.apache.thrift.ProcessFunction<I, ? extends org.apache.thrift.TBase>>()));
}
- private Iface iface_;
+ protected Processor(I iface, Map<String, org.apache.thrift.ProcessFunction<I, ? extends org.apache.thrift.TBase>> processMap) {
+ super(iface, getProcessMap(processMap));
+ }
- public boolean process(TProtocol iprot, TProtocol oprot) throws TException
- {
- TMessage msg = iprot.readMessageBegin();
- ProcessFunction fn = processMap_.get(msg.name);
- if (fn == null) {
- TProtocolUtil.skip(iprot, TType.STRUCT);
- iprot.readMessageEnd();
- TApplicationException x = new TApplicationException(TApplicationException.UNKNOWN_METHOD, "Invalid method name: '"+msg.name+"'");
- oprot.writeMessageBegin(new TMessage(msg.name, TMessageType.EXCEPTION, msg.seqid));
- x.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
- return true;
- }
- fn.process(msg.seqid, iprot, oprot);
- return true;
+ private static <I extends Iface> Map<String, org.apache.thrift.ProcessFunction<I, ? extends org.apache.thrift.TBase>> getProcessMap(Map<String, org.apache.thrift.ProcessFunction<I, ? extends org.apache.thrift.TBase>> processMap) {
+ processMap.put("execute", new execute());
+ processMap.put("fetchOne", new fetchOne());
+ processMap.put("fetchN", new fetchN());
+ processMap.put("fetchAll", new fetchAll());
+ processMap.put("getSchema", new getSchema());
+ processMap.put("getThriftSchema", new getThriftSchema());
+ processMap.put("getClusterStatus", new getClusterStatus());
+ processMap.put("getQueryPlan", new getQueryPlan());
+ processMap.put("clean", new clean());
+ return processMap;
}
- private class execute implements ProcessFunction {
- public void process(int seqid, TProtocol iprot, TProtocol oprot) throws TException
- {
- execute_args args = new execute_args();
- try {
- args.read(iprot);
- } catch (TProtocolException e) {
- iprot.readMessageEnd();
- TApplicationException x = new TApplicationException(TApplicationException.PROTOCOL_ERROR, e.getMessage());
- oprot.writeMessageBegin(new TMessage("execute", TMessageType.EXCEPTION, seqid));
- x.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
- return;
- }
- iprot.readMessageEnd();
+ private static class execute<I extends Iface> extends org.apache.thrift.ProcessFunction<I, execute_args> {
+ public execute() {
+ super("execute");
+ }
+
+ protected execute_args getEmptyArgsInstance() {
+ return new execute_args();
+ }
+
+ protected execute_result getResult(I iface, execute_args args) throws org.apache.thrift.TException {
execute_result result = new execute_result();
try {
- iface_.execute(args.query);
+ iface.execute(args.query);
} catch (HiveServerException ex) {
result.ex = ex;
- } catch (Throwable th) {
- LOGGER.error("Internal error processing execute", th);
- TApplicationException x = new TApplicationException(TApplicationException.INTERNAL_ERROR, "Internal error processing execute");
- oprot.writeMessageBegin(new TMessage("execute", TMessageType.EXCEPTION, seqid));
- x.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
- return;
}
- oprot.writeMessageBegin(new TMessage("execute", TMessageType.REPLY, seqid));
- result.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
+ return result;
}
-
}
- private class fetchOne implements ProcessFunction {
- public void process(int seqid, TProtocol iprot, TProtocol oprot) throws TException
- {
- fetchOne_args args = new fetchOne_args();
- try {
- args.read(iprot);
- } catch (TProtocolException e) {
- iprot.readMessageEnd();
- TApplicationException x = new TApplicationException(TApplicationException.PROTOCOL_ERROR, e.getMessage());
- oprot.writeMessageBegin(new TMessage("fetchOne", TMessageType.EXCEPTION, seqid));
- x.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
- return;
- }
- iprot.readMessageEnd();
+ private static class fetchOne<I extends Iface> extends org.apache.thrift.ProcessFunction<I, fetchOne_args> {
+ public fetchOne() {
+ super("fetchOne");
+ }
+
+ protected fetchOne_args getEmptyArgsInstance() {
+ return new fetchOne_args();
+ }
+
+ protected fetchOne_result getResult(I iface, fetchOne_args args) throws org.apache.thrift.TException {
fetchOne_result result = new fetchOne_result();
try {
- result.success = iface_.fetchOne();
+ result.success = iface.fetchOne();
} catch (HiveServerException ex) {
result.ex = ex;
- } catch (Throwable th) {
- LOGGER.error("Internal error processing fetchOne", th);
- TApplicationException x = new TApplicationException(TApplicationException.INTERNAL_ERROR, "Internal error processing fetchOne");
- oprot.writeMessageBegin(new TMessage("fetchOne", TMessageType.EXCEPTION, seqid));
- x.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
- return;
}
- oprot.writeMessageBegin(new TMessage("fetchOne", TMessageType.REPLY, seqid));
- result.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
+ return result;
}
-
}
- private class fetchN implements ProcessFunction {
- public void process(int seqid, TProtocol iprot, TProtocol oprot) throws TException
- {
- fetchN_args args = new fetchN_args();
- try {
- args.read(iprot);
- } catch (TProtocolException e) {
- iprot.readMessageEnd();
- TApplicationException x = new TApplicationException(TApplicationException.PROTOCOL_ERROR, e.getMessage());
- oprot.writeMessageBegin(new TMessage("fetchN", TMessageType.EXCEPTION, seqid));
- x.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
- return;
- }
- iprot.readMessageEnd();
+ private static class fetchN<I extends Iface> extends org.apache.thrift.ProcessFunction<I, fetchN_args> {
+ public fetchN() {
+ super("fetchN");
+ }
+
+ protected fetchN_args getEmptyArgsInstance() {
+ return new fetchN_args();
+ }
+
+ protected fetchN_result getResult(I iface, fetchN_args args) throws org.apache.thrift.TException {
fetchN_result result = new fetchN_result();
try {
- result.success = iface_.fetchN(args.numRows);
+ result.success = iface.fetchN(args.numRows);
} catch (HiveServerException ex) {
result.ex = ex;
- } catch (Throwable th) {
- LOGGER.error("Internal error processing fetchN", th);
- TApplicationException x = new TApplicationException(TApplicationException.INTERNAL_ERROR, "Internal error processing fetchN");
- oprot.writeMessageBegin(new TMessage("fetchN", TMessageType.EXCEPTION, seqid));
- x.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
- return;
}
- oprot.writeMessageBegin(new TMessage("fetchN", TMessageType.REPLY, seqid));
- result.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
+ return result;
}
-
}
- private class fetchAll implements ProcessFunction {
- public void process(int seqid, TProtocol iprot, TProtocol oprot) throws TException
- {
- fetchAll_args args = new fetchAll_args();
- try {
- args.read(iprot);
- } catch (TProtocolException e) {
- iprot.readMessageEnd();
- TApplicationException x = new TApplicationException(TApplicationException.PROTOCOL_ERROR, e.getMessage());
- oprot.writeMessageBegin(new TMessage("fetchAll", TMessageType.EXCEPTION, seqid));
- x.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
- return;
- }
- iprot.readMessageEnd();
+ private static class fetchAll<I extends Iface> extends org.apache.thrift.ProcessFunction<I, fetchAll_args> {
+ public fetchAll() {
+ super("fetchAll");
+ }
+
+ protected fetchAll_args getEmptyArgsInstance() {
+ return new fetchAll_args();
+ }
+
+ protected fetchAll_result getResult(I iface, fetchAll_args args) throws org.apache.thrift.TException {
fetchAll_result result = new fetchAll_result();
try {
- result.success = iface_.fetchAll();
+ result.success = iface.fetchAll();
} catch (HiveServerException ex) {
result.ex = ex;
- } catch (Throwable th) {
- LOGGER.error("Internal error processing fetchAll", th);
- TApplicationException x = new TApplicationException(TApplicationException.INTERNAL_ERROR, "Internal error processing fetchAll");
- oprot.writeMessageBegin(new TMessage("fetchAll", TMessageType.EXCEPTION, seqid));
- x.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
- return;
}
- oprot.writeMessageBegin(new TMessage("fetchAll", TMessageType.REPLY, seqid));
- result.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
+ return result;
}
-
}
- private class getSchema implements ProcessFunction {
- public void process(int seqid, TProtocol iprot, TProtocol oprot) throws TException
- {
- getSchema_args args = new getSchema_args();
- try {
- args.read(iprot);
- } catch (TProtocolException e) {
- iprot.readMessageEnd();
- TApplicationException x = new TApplicationException(TApplicationException.PROTOCOL_ERROR, e.getMessage());
- oprot.writeMessageBegin(new TMessage("getSchema", TMessageType.EXCEPTION, seqid));
- x.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
- return;
- }
- iprot.readMessageEnd();
+ private static class getSchema<I extends Iface> extends org.apache.thrift.ProcessFunction<I, getSchema_args> {
+ public getSchema() {
+ super("getSchema");
+ }
+
+ protected getSchema_args getEmptyArgsInstance() {
+ return new getSchema_args();
+ }
+
+ protected getSchema_result getResult(I iface, getSchema_args args) throws org.apache.thrift.TException {
getSchema_result result = new getSchema_result();
try {
- result.success = iface_.getSchema();
+ result.success = iface.getSchema();
} catch (HiveServerException ex) {
result.ex = ex;
- } catch (Throwable th) {
- LOGGER.error("Internal error processing getSchema", th);
- TApplicationException x = new TApplicationException(TApplicationException.INTERNAL_ERROR, "Internal error processing getSchema");
- oprot.writeMessageBegin(new TMessage("getSchema", TMessageType.EXCEPTION, seqid));
- x.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
- return;
}
- oprot.writeMessageBegin(new TMessage("getSchema", TMessageType.REPLY, seqid));
- result.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
+ return result;
}
-
}
- private class getThriftSchema implements ProcessFunction {
- public void process(int seqid, TProtocol iprot, TProtocol oprot) throws TException
- {
- getThriftSchema_args args = new getThriftSchema_args();
- try {
- args.read(iprot);
- } catch (TProtocolException e) {
- iprot.readMessageEnd();
- TApplicationException x = new TApplicationException(TApplicationException.PROTOCOL_ERROR, e.getMessage());
- oprot.writeMessageBegin(new TMessage("getThriftSchema", TMessageType.EXCEPTION, seqid));
- x.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
- return;
- }
- iprot.readMessageEnd();
+ private static class getThriftSchema<I extends Iface> extends org.apache.thrift.ProcessFunction<I, getThriftSchema_args> {
+ public getThriftSchema() {
+ super("getThriftSchema");
+ }
+
+ protected getThriftSchema_args getEmptyArgsInstance() {
+ return new getThriftSchema_args();
+ }
+
+ protected getThriftSchema_result getResult(I iface, getThriftSchema_args args) throws org.apache.thrift.TException {
getThriftSchema_result result = new getThriftSchema_result();
try {
- result.success = iface_.getThriftSchema();
+ result.success = iface.getThriftSchema();
} catch (HiveServerException ex) {
result.ex = ex;
- } catch (Throwable th) {
- LOGGER.error("Internal error processing getThriftSchema", th);
- TApplicationException x = new TApplicationException(TApplicationException.INTERNAL_ERROR, "Internal error processing getThriftSchema");
- oprot.writeMessageBegin(new TMessage("getThriftSchema", TMessageType.EXCEPTION, seqid));
- x.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
- return;
}
- oprot.writeMessageBegin(new TMessage("getThriftSchema", TMessageType.REPLY, seqid));
- result.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
+ return result;
}
-
}
- private class getClusterStatus implements ProcessFunction {
- public void process(int seqid, TProtocol iprot, TProtocol oprot) throws TException
- {
- getClusterStatus_args args = new getClusterStatus_args();
- try {
- args.read(iprot);
- } catch (TProtocolException e) {
- iprot.readMessageEnd();
- TApplicationException x = new TApplicationException(TApplicationException.PROTOCOL_ERROR, e.getMessage());
- oprot.writeMessageBegin(new TMessage("getClusterStatus", TMessageType.EXCEPTION, seqid));
- x.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
- return;
- }
- iprot.readMessageEnd();
+ private static class getClusterStatus<I extends Iface> extends org.apache.thrift.ProcessFunction<I, getClusterStatus_args> {
+ public getClusterStatus() {
+ super("getClusterStatus");
+ }
+
+ protected getClusterStatus_args getEmptyArgsInstance() {
+ return new getClusterStatus_args();
+ }
+
+ protected getClusterStatus_result getResult(I iface, getClusterStatus_args args) throws org.apache.thrift.TException {
getClusterStatus_result result = new getClusterStatus_result();
try {
- result.success = iface_.getClusterStatus();
+ result.success = iface.getClusterStatus();
} catch (HiveServerException ex) {
result.ex = ex;
- } catch (Throwable th) {
- LOGGER.error("Internal error processing getClusterStatus", th);
- TApplicationException x = new TApplicationException(TApplicationException.INTERNAL_ERROR, "Internal error processing getClusterStatus");
- oprot.writeMessageBegin(new TMessage("getClusterStatus", TMessageType.EXCEPTION, seqid));
- x.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
- return;
}
- oprot.writeMessageBegin(new TMessage("getClusterStatus", TMessageType.REPLY, seqid));
- result.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
+ return result;
}
-
}
- private class getQueryPlan implements ProcessFunction {
- public void process(int seqid, TProtocol iprot, TProtocol oprot) throws TException
- {
- getQueryPlan_args args = new getQueryPlan_args();
- try {
- args.read(iprot);
- } catch (TProtocolException e) {
- iprot.readMessageEnd();
- TApplicationException x = new TApplicationException(TApplicationException.PROTOCOL_ERROR, e.getMessage());
- oprot.writeMessageBegin(new TMessage("getQueryPlan", TMessageType.EXCEPTION, seqid));
- x.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
- return;
- }
- iprot.readMessageEnd();
+ private static class getQueryPlan<I extends Iface> extends org.apache.thrift.ProcessFunction<I, getQueryPlan_args> {
+ public getQueryPlan() {
+ super("getQueryPlan");
+ }
+
+ protected getQueryPlan_args getEmptyArgsInstance() {
+ return new getQueryPlan_args();
+ }
+
+ protected getQueryPlan_result getResult(I iface, getQueryPlan_args args) throws org.apache.thrift.TException {
getQueryPlan_result result = new getQueryPlan_result();
try {
- result.success = iface_.getQueryPlan();
+ result.success = iface.getQueryPlan();
} catch (HiveServerException ex) {
result.ex = ex;
- } catch (Throwable th) {
- LOGGER.error("Internal error processing getQueryPlan", th);
- TApplicationException x = new TApplicationException(TApplicationException.INTERNAL_ERROR, "Internal error processing getQueryPlan");
- oprot.writeMessageBegin(new TMessage("getQueryPlan", TMessageType.EXCEPTION, seqid));
- x.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
- return;
}
- oprot.writeMessageBegin(new TMessage("getQueryPlan", TMessageType.REPLY, seqid));
- result.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
+ return result;
}
-
}
- private class clean implements ProcessFunction {
- public void process(int seqid, TProtocol iprot, TProtocol oprot) throws TException
- {
- clean_args args = new clean_args();
- try {
- args.read(iprot);
- } catch (TProtocolException e) {
- iprot.readMessageEnd();
- TApplicationException x = new TApplicationException(TApplicationException.PROTOCOL_ERROR, e.getMessage());
- oprot.writeMessageBegin(new TMessage("clean", TMessageType.EXCEPTION, seqid));
- x.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
- return;
- }
- iprot.readMessageEnd();
- clean_result result = new clean_result();
- iface_.clean();
- oprot.writeMessageBegin(new TMessage("clean", TMessageType.REPLY, seqid));
- result.write(oprot);
- oprot.writeMessageEnd();
- oprot.getTransport().flush();
+ private static class clean<I extends Iface> extends org.apache.thrift.ProcessFunction<I, clean_args> {
+ public clean() {
+ super("clean");
+ }
+
+ protected clean_args getEmptyArgsInstance() {
+ return new clean_args();
}
+ protected clean_result getResult(I iface, clean_args args) throws org.apache.thrift.TException {
+ clean_result result = new clean_result();
+ iface.clean();
+ return result;
+ }
}
}
- public static class execute_args implements TBase<execute_args, execute_args._Fields>, java.io.Serializable, Cloneable {
- private static final TStruct STRUCT_DESC = new TStruct("execute_args");
+ public static class execute_args implements org.apache.thrift.TBase<execute_args, execute_args._Fields>, java.io.Serializable, Cloneable {
+ private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("execute_args");
- private static final TField QUERY_FIELD_DESC = new TField("query", TType.STRING, (short)1);
+ private static final org.apache.thrift.protocol.TField QUERY_FIELD_DESC = new org.apache.thrift.protocol.TField("query", org.apache.thrift.protocol.TType.STRING, (short)1);
- private String query;
+ private String query; // required
/** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */
- public enum _Fields implements TFieldIdEnum {
+ public enum _Fields implements org.apache.thrift.TFieldIdEnum {
QUERY((short)1, "query");
private static final Map<String, _Fields> byName = new HashMap<String, _Fields>();
@@ -1142,13 +859,13 @@ public class ThriftHive {
// isset id assignments
- public static final Map<_Fields, FieldMetaData> metaDataMap;
+ public static final Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap;
static {
- Map<_Fields, FieldMetaData> tmpMap = new EnumMap<_Fields, FieldMetaData>(_Fields.class);
- tmpMap.put(_Fields.QUERY, new FieldMetaData("query", TFieldRequirementType.DEFAULT,
- new FieldValueMetaData(TType.STRING)));
+ Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class);
+ tmpMap.put(_Fields.QUERY, new org.apache.thrift.meta_data.FieldMetaData("query", org.apache.thrift.TFieldRequirementType.DEFAULT,
+ new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING)));
metaDataMap = Collections.unmodifiableMap(tmpMap);
- FieldMetaData.addStructMetaDataMap(execute_args.class, metaDataMap);
+ org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(execute_args.class, metaDataMap);
}
public execute_args() {
@@ -1191,7 +908,7 @@ public class ThriftHive {
this.query = null;
}
- /** Returns true if field query is set (has been asigned a value) and false otherwise */
+ /** Returns true if field query is set (has been assigned a value) and false otherwise */
public boolean isSetQuery() {
return this.query != null;
}
@@ -1224,7 +941,7 @@ public class ThriftHive {
throw new IllegalStateException();
}
- /** Returns true if field corresponding to fieldID is set (has been asigned a value) and false otherwise */
+ /** Returns true if field corresponding to fieldID is set (has been assigned a value) and false otherwise */
public boolean isSet(_Fields field) {
if (field == null) {
throw new IllegalArgumentException();
@@ -1280,7 +997,7 @@ public class ThriftHive {
return lastComparison;
}
if (isSetQuery()) {
- lastComparison = TBaseHelper.compareTo(this.query, typedOther.query);
+ lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.query, typedOther.query);
if (lastComparison != 0) {
return lastComparison;
}
@@ -1292,25 +1009,25 @@ public class ThriftHive {
return _Fields.findByThriftId(fieldId);
}
- public void read(TProtocol iprot) throws TException {
- TField field;
+ public void read(org.apache.thrift.protocol.TProtocol iprot) throws org.apache.thrift.TException {
+ org.apache.thrift.protocol.TField field;
iprot.readStructBegin();
while (true)
{
field = iprot.readFieldBegin();
- if (field.type == TType.STOP) {
+ if (field.type == org.apache.thrift.protocol.TType.STOP) {
break;
}
switch (field.id) {
case 1: // QUERY
- if (field.type == TType.STRING) {
+ if (field.type == org.apache.thrift.protocol.TType.STRING) {
this.query = iprot.readString();
} else {
- TProtocolUtil.skip(iprot, field.type);
+ org.apache.thrift.protocol.TProtocolUtil.skip(iprot, field.type);
}
break;
default:
- TProtocolUtil.skip(iprot, field.type);
+ org.apache.thrift.protocol.TProtocolUtil.skip(iprot, field.type);
}
iprot.readFieldEnd();
}
@@ -1318,7 +1035,7 @@ public class ThriftHive {
validate();
}
- public void write(TProtocol oprot) throws TException {
+ public void write(org.apache.thrift.protocol.TProtocol oprot) throws org.apache.thrift.TException {
validate();
oprot.writeStructBegin(STRUCT_DESC);
@@ -1347,21 +1064,37 @@ public class ThriftHive {
return sb.toString();
}
- public void validate() throws TException {
+ public void validate() throws org.apache.thrift.TException {
// check for required fields
}
+ private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException {
+ try {
+ write(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(out)));
+ } catch (org.apache.thrift.TException te) {
+ throw new java.io.IOException(te);
+ }
+ }
+
+ private void readObject(java.io.ObjectInputStream in) throws java.io.IOException, ClassNotFoundException {
+ try {
+ read(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(in)));
+ } catch (org.apache.thrift.TException te) {
+ throw new java.io.IOException(te);
+ }
+ }
+
}
- public static class execute_result implements TBase<execute_result, execute_result._Fields>, java.io.Serializable, Cloneable {
- private static final TStruct STRUCT_DESC = new TStruct("execute_result");
+ public static class execute_result implements org.apache.thrift.TBase<execute_result, execute_result._Fields>, java.io.Serializable, Cloneable {
+ private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("execute_result");
- private static final TField EX_FIELD_DESC = new TField("ex", TType.STRUCT, (short)1);
+ private static final org.apache.thrift.protocol.TField EX_FIELD_DESC = new org.apache.thrift.protocol.TField("ex", org.apache.thrift.protocol.TType.STRUCT, (short)1);
- private HiveServerException ex;
+ private HiveServerException ex; // required
/** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */
- public enum _Fields implements TFieldIdEnum {
+ public enum _Fields implements org.apache.thrift.TFieldIdEnum {
EX((short)1, "ex");
private static final Map<String, _Fields> byName = new HashMap<String, _Fields>();
@@ -1420,13 +1153,13 @@ public class ThriftHive {
// isset id assignments
- public static final Map<_Fields, FieldMetaData> metaDataMap;
+ public static final Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap;
static {
- Map<_Fields, FieldMetaData> tmpMap = new EnumMap<_Fields, FieldMetaData>(_Fields.class);
- tmpMap.put(_Fields.EX, new FieldMetaData("ex", TFieldRequirementType.DEFAULT,
- new FieldValueMetaData(TType.STRUCT)));
+ Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class);
+ tmpMap.put(_Fields.EX, new org.apache.thrift.meta_data.FieldMetaData("ex", org.apache.thrift.TFieldRequirementType.DEFAULT,
+ new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRUCT)));
metaDataMap = Collections.unmodifiableMap(tmpMap);
- FieldMetaData.addStructMetaDataMap(execute_result.class, metaDataMap);
+ org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(execute_result.class, metaDataMap);
}
public execute_result() {
@@ -1469,7 +1202,7 @@ public class ThriftHive {
this.ex = null;
}
- /** Returns true if field ex is set (has been asigned a value) and false otherwise */
+ /** Returns true if field ex is set (has been assigned a value) and false otherwise */
public boolean isSetEx() {
return this.ex != null;
}
@@ -1502,7 +1235,7 @@ public class ThriftHive {
throw new IllegalStateException();
}
- /** Returns true if field corresponding to fieldID is set (has been asigned a value) and false otherwise */
+ /** Returns true if field corresponding to fieldID is set (has been assigned a value) and false otherwise */
public boolean isSet(_Fields field) {
if (field == null) {
throw new IllegalArgumentException();
@@ -1558,7 +1291,7 @@ public class ThriftHive {
return lastComparison;
}
if (isSetEx()) {
- lastComparison = TBaseHelper.compareTo(this.ex, typedOther.ex);
+ lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.ex, typedOther.ex);
if (lastComparison != 0) {
return lastComparison;
}
@@ -1570,26 +1303,26 @@ public class ThriftHive {
return _Fields.findByThriftId(fieldId);
}
- public void read(TProtocol iprot) throws TException {
- TField field;
+ public void read(org.apache.thrift.protocol.TProtocol iprot) throws org.apache.thrift.TException {
+ org.apache.thrift.protocol.TField field;
iprot.readStructBegin();
while (true)
{
field = iprot.readFieldBegin();
- if (field.type == TType.STOP) {
+ if (field.type == org.apache.thrift.protocol.TType.STOP) {
break;
}
switch (field.id) {
case 1: // EX
- if (field.type == TType.STRUCT) {
+ if (field.type == org.apache.thrift.protocol.TType.STRUCT) {
this.ex = new HiveServerException();
this.ex.read(iprot);
} else {
- TProtocolUtil.skip(iprot, field.type);
+ org.apache.thrift.protocol.TProtocolUtil.skip(iprot, field.type);
}
break;
default:
- TProtocolUtil.skip(iprot, field.type);
+ org.apache.thrift.protocol.TProtocolUtil.skip(iprot, field.type);
}
iprot.readFieldEnd();
}
@@ -1597,7 +1330,7 @@ public class ThriftHive {
validate();
}
- public void write(TProtocol oprot) throws TException {
+ public void write(org.apache.thrift.protocol.TProtocol oprot) throws org.apache.thrift.TException {
oprot.writeStructBegin(STRUCT_DESC);
if (this.isSetEx()) {
@@ -1625,19 +1358,35 @@ public class ThriftHive {
return sb.toString();
}
- public void validate() throws TException {
+ public void validate() throws org.apache.thrift.TException {
// check for required fields
}
+ private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException {
+ try {
+ write(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(out)));
+ } catch (org.apache.thrift.TException te) {
+ throw new java.io.IOException(te);
+ }
+ }
+
+ private void readObject(java.io.ObjectInputStream in) throws java.io.IOException, ClassNotFoundException {
+ try {
+ read(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(in)));
+ } catch (org.apache.thrift.TException te) {
+ throw new java.io.IOException(te);
+ }
+ }
+
}
- public static class fetchOne_args implements TBase<fetchOne_args, fetchOne_args._Fields>, java.io.Serializable, Cloneable {
- private static final TStruct STRUCT_DESC = new TStruct("fetchOne_args");
+ public static class fetchOne_args implements org.apache.thrift.TBase<fetchOne_args, fetchOne_args._Fields>, java.io.Serializable, Cloneable {
+ private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("fetchOne_args");
/** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */
- public enum _Fields implements TFieldIdEnum {
+ public enum _Fields implements org.apache.thrift.TFieldIdEnum {
;
private static final Map<String, _Fields> byName = new HashMap<String, _Fields>();
@@ -1691,11 +1440,11 @@ public class ThriftHive {
return _fieldName;
}
}
- public static final Map<_Fields, FieldMetaData> metaDataMap;
+ public static final Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap;
static {
- Map<_Fields, FieldMetaData> tmpMap = new EnumMap<_Fields, FieldMetaData>(_Fields.class);
+ Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class);
metaDataMap = Collections.unmodifiableMap(tmpMap);
- FieldMetaData.addStructMetaDataMap(fetchOne_args.class, metaDataMap);
+ org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(fetchOne_args.class, metaDataMap);
}
public fetchOne_args() {
@@ -1726,7 +1475,7 @@ public class ThriftHive {
throw new IllegalStateException();
}
- /** Returns true if field corresponding to fieldID is set (has been asigned a value) and false otherwise */
+ /** Returns true if field corresponding to fieldID is set (has been assigned a value) and false otherwise */
public boolean isSet(_Fields field) {
if (field == null) {
throw new IllegalArgumentException();
@@ -1773,18 +1522,18 @@ public class ThriftHive {
return _Fields.findByThriftId(fieldId);
}
- public void read(TProtocol iprot) throws TException {
- TField field;
+ public void read(org.apache.thrift.protocol.TProtocol iprot) throws org.apache.thrift.TException {
+ org.apache.thrift.protocol.TField field;
iprot.readStructBegin();
while (true)
{
field = iprot.readFieldBegin();
- if (field.type == TType.STOP) {
+ if (field.type == org.apache.thrift.protocol.TType.STOP) {
break;
}
switch (field.id) {
default:
- TProtocolUtil.skip(iprot, field.type);
+ org.apache.thrift.protocol.TProtocolUtil.skip(iprot, field.type);
}
iprot.readFieldEnd();
}
@@ -1792,7 +1541,7 @@ public class ThriftHive {
validate();
}
- public void write(TProtocol oprot) throws TException {
+ public void write(org.apache.thrift.protocol.TProtocol oprot) throws org.apache.thrift.TException {
validate();
oprot.writeStructBegin(STRUCT_DESC);
@@ -1809,23 +1558,39 @@ public class ThriftHive {
return sb.toString();
}
- public void validate() throws TException {
+ public void validate() throws org.apache.thrift.TException {
// check for required fields
}
+ private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException {
+ try {
+ write(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(out)));
+ } catch (org.apache.thrift.TException te) {
+ throw new java.io.IOException(te);
+ }
+ }
+
+ private void readObject(java.io.ObjectInputStream in) throws java.io.IOException, ClassNotFoundException {
+ try {
+ read(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(in)));
+ } catch (org.apache.thrift.TException te) {
+ throw new java.io.IOException(te);
+ }
+ }
+
}
- public static class fetchOne_result implements TBase<fetchOne_result, fetchOne_result._Fields>, java.io.Serializable, Cloneable {
- private static final TStruct STRUCT_DESC = new TStruct("fetchOne_result");
+ public static class fetchOne_result implements org.apache.thrift.TBase<fetchOne_result, fetchOne_result._Fields>, java.io.Serializable, Cloneable {
+ private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("fetchOne_result");
- private static final TField SUCCESS_FIELD_DESC = new TField("success", TType.STRING, (short)0);
- private static final TField EX_FIELD_DESC = new TField("ex", TType.STRUCT, (short)1);
+ private static final org.apache.thrift.protocol.TField SUCCESS_FIELD_DESC = new org.apache.thrift.protocol.TField("success", org.apache.thrift.protocol.TType.STRING, (short)0);
+ private static final org.apache.thrift.protocol.TField EX_FIELD_DESC = new org.apache.thrift.protocol.TField("ex", org.apache.thrift.protocol.TType.STRUCT, (short)1);
- private String success;
- private HiveServerException ex;
+ private String success; // required
+ private HiveServerException ex; // required
/** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */
- public enum _Fields implements TFieldIdEnum {
+ public enum _Fields implements org.apache.thrift.TFieldIdEnum {
SUCCESS((short)0, "success"),
EX((short)1, "ex");
@@ -1887,15 +1652,15 @@ public class ThriftHive {
// isset id assignments
- public static final Map<_Fields, FieldMetaData> metaDataMap;
+ public static final Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap;
static {
- Map<_Fields, FieldMetaData> tmpMap = new EnumMap<_Fields, FieldMetaData>(_Fields.class);
- tmpMap.put(_Fields.SUCCESS, new FieldMetaData("success", TFieldRequirementType.DEFAULT,
- new FieldValueMetaData(TType.STRING)));
- tmpMap.put(_Fields.EX, new FieldMetaData("ex", TFieldRequirementType.DEFAULT,
- new FieldValueMetaData(TType.STRUCT)));
+ Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class);
+ tmpMap.put(_Fields.SUCCESS, new org.apache.thrift.meta_data.FieldMetaData("success", org.apache.thrift.TFieldRequirementType.DEFAULT,
+ new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING)));
+ tmpMap.put(_Fields.EX, new org.apache.thrift.meta_data.FieldMetaData("ex", org.apache.thrift.TFieldRequirementType.DEFAULT,
+ new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRUCT)));
metaDataMap = Collections.unmodifiableMap(tmpMap);
[... 1991 lines stripped ...]