You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@drill.apache.org by vo...@apache.org on 2019/06/18 18:22:49 UTC

[drill] branch master updated (de0aec7 -> f3f7dbd)

This is an automated email from the ASF dual-hosted git repository.

volodymyr pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/drill.git.


    from de0aec7  DRILL-7251: Read Hive array w/o nulls
     new 4eb7263  DRILL-7294: Prevent generating java beans using protostuff to avoid overriding classes with the same simple name declared as nested in the proto files
     new 24ec980  DRILL-7294: Regenerate protobufs
     new f3f7dbd  DRILL-7292: Remove V1 and V2 text readers

The 3 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.


Summary of changes:
 .../apache/drill/exec/store/jdbc/JdbcSubScan.java  |    4 +-
 .../java/org/apache/drill/exec/ExecConstants.java  |   19 +-
 .../exec/coord/store/TransientStoreConfig.java     |    2 +-
 .../impl/scan/columns/ColumnsArrayParser.java      |    2 +-
 .../impl/statistics/AvgWidthMergedStatistic.java   |    7 +-
 .../impl/statistics/NDVMergedStatistic.java        |    7 +-
 .../drill/exec/serialization/ProtoSerializer.java  |    4 +-
 .../exec/store/easy/text/TextFormatPlugin.java     |   57 +-
 .../text/compliant/CompliantTextRecordReader.java  |  256 ----
 .../easy/text/compliant/FieldVarCharOutput.java    |  234 ----
 .../store/easy/text/compliant/HeaderBuilder.java   |  263 ----
 .../easy/text/compliant/RepeatedVarCharOutput.java |  346 -----
 .../compliant/StreamFinishedPseudoException.java   |   29 -
 .../exec/store/easy/text/compliant/TextInput.java  |  365 ------
 .../exec/store/easy/text/compliant/TextOutput.java |   87 --
 .../easy/text/compliant/TextParsingContext.java    |  122 --
 .../easy/text/compliant/TextParsingSettings.java   |  291 -----
 .../exec/store/easy/text/compliant/TextReader.java |  511 --------
 .../store/easy/text/compliant/package-info.java    |   27 -
 .../{compliant/v3 => reader}/BaseFieldOutput.java  |    2 +-
 .../v3 => reader}/CompliantTextBatchReader.java    |    6 +-
 .../v3 => reader}/FieldVarCharOutput.java          |    2 +-
 .../{compliant/v3 => reader}/HeaderBuilder.java    |    2 +-
 .../v3 => reader}/RepeatedVarCharOutput.java       |    2 +-
 .../StreamFinishedPseudoException.java             |    9 +-
 .../text/{compliant/v3 => reader}/TextInput.java   |    6 +-
 .../text/{compliant/v3 => reader}/TextOutput.java  |    2 +-
 .../v3 => reader}/TextParsingContext.java          |    2 +-
 .../TextParsingSettings.java}                      |    7 +-
 .../text/{compliant/v3 => reader}/TextReader.java  |   27 +-
 .../{compliant/v3 => reader}/package-info.java     |    2 +-
 .../text/writer/TextRecordWriter.java}             |   14 +-
 .../exec/store/sys/PersistentStoreConfig.java      |    2 +-
 .../exec/store/text/DrillTextRecordReader.java     |  244 ----
 .../org/apache/drill/TestSelectWithOption.java     |   13 +-
 .../drill/exec/store/TestImplicitFileColumns.java  |   16 +-
 .../store/easy/text/compliant/BaseCsvTest.java     |   10 -
 .../text/compliant}/TestCsvHeader.java             |    2 +-
 .../easy/text/compliant/TestCsvIgnoreHeaders.java  |   25 +-
 .../text/compliant/TestCsvTableProperties.java     |   67 +
 .../easy/text/compliant/TestCsvWithHeaders.java    |  846 ++++--------
 .../easy/text/compliant/TestCsvWithSchema.java     |  188 +--
 .../easy/text/compliant/TestCsvWithoutHeaders.java |  344 ++---
 .../easy/text/compliant/TestHeaderBuilder.java     |    2 +-
 .../easy/text/compliant/TestPartitionRace.java     |  227 +---
 .../text/compliant}/TestTextColumn.java            |   15 +-
 .../text/compliant/TestTextReader.java}            |   11 +-
 .../text/compliant}/TextRecordReaderTest.java      |    6 +-
 .../src/test/resources/store/text/data/letters.csv |    9 +-
 exec/jdbc-all/pom.xml                              |    4 +-
 protocol/pom.xml                                   |   35 +-
 .../org/apache/drill/common/types/DataMode.java    |   51 -
 .../org/apache/drill/common/types/MajorType.java   |  305 -----
 .../org/apache/drill/common/types/MinorType.java   |  123 --
 .../drill/common/types/SchemaTypeProtos.java       |   12 +-
 .../apache/drill/exec/proto/SchemaBitControl.java  |  120 +-
 .../org/apache/drill/exec/proto/SchemaBitData.java |   48 +-
 .../drill/exec/proto/SchemaCoordinationProtos.java |   36 +-
 .../apache/drill/exec/proto/SchemaExecProtos.java  |   24 +-
 .../drill/exec/proto/SchemaGeneralRPCProtos.java   |   36 +-
 .../drill/exec/proto/SchemaUserBitShared.java      |  264 ++--
 .../apache/drill/exec/proto/SchemaUserProtos.java  |  372 +++---
 .../org/apache/drill/exec/proto/beans/Ack.java     |  163 ---
 .../drill/exec/proto/beans/BitClientHandshake.java |  185 ---
 .../exec/proto/beans/BitControlHandshake.java      |  241 ----
 .../drill/exec/proto/beans/BitServerHandshake.java |  195 ---
 .../apache/drill/exec/proto/beans/BitStatus.java   |  175 ---
 .../drill/exec/proto/beans/BitToUserHandshake.java |  359 ------
 .../drill/exec/proto/beans/CatalogMetadata.java    |  207 ---
 .../drill/exec/proto/beans/CollateSupport.java     |   49 -
 .../apache/drill/exec/proto/beans/Collector.java   |  239 ----
 .../drill/exec/proto/beans/ColumnMetadata.java     |  515 --------
 .../exec/proto/beans/ColumnSearchability.java      |   55 -
 .../drill/exec/proto/beans/ColumnUpdatability.java |   51 -
 .../drill/exec/proto/beans/CompleteRpcMessage.java |  210 ---
 .../drill/exec/proto/beans/ConvertSupport.java     |  199 ---
 .../drill/exec/proto/beans/CoreOperatorType.java   |  171 ---
 .../exec/proto/beans/CorrelationNamesSupport.java  |   51 -
 .../proto/beans/CreatePreparedStatementReq.java    |  163 ---
 .../proto/beans/CreatePreparedStatementResp.java   |  211 ---
 .../drill/exec/proto/beans/CustomMessage.java      |  186 ---
 .../exec/proto/beans/DateTimeLiteralsSupport.java  |   79 --
 .../drill/exec/proto/beans/DrillPBError.java       |  341 -----
 .../exec/proto/beans/DrillServiceInstance.java     |  209 ---
 .../drill/exec/proto/beans/DrillbitEndpoint.java   |  351 -----
 .../drill/exec/proto/beans/ExceptionWrapper.java   |  243 ----
 .../drill/exec/proto/beans/FinishedReceiver.java   |  189 ---
 .../drill/exec/proto/beans/FragmentHandle.java     |  233 ----
 .../exec/proto/beans/FragmentRecordBatch.java      |  309 -----
 .../drill/exec/proto/beans/FragmentState.java      |   59 -
 .../drill/exec/proto/beans/FragmentStatus.java     |  189 ---
 .../drill/exec/proto/beans/GetCatalogsReq.java     |  165 ---
 .../drill/exec/proto/beans/GetCatalogsResp.java    |  221 ----
 .../drill/exec/proto/beans/GetColumnsReq.java      |  237 ----
 .../drill/exec/proto/beans/GetColumnsResp.java     |  221 ----
 .../exec/proto/beans/GetQueryPlanFragments.java    |  218 ----
 .../drill/exec/proto/beans/GetSchemasReq.java      |  189 ---
 .../drill/exec/proto/beans/GetSchemasResp.java     |  221 ----
 .../drill/exec/proto/beans/GetServerMetaResp.java  |  211 ---
 .../drill/exec/proto/beans/GetTablesReq.java       |  245 ----
 .../drill/exec/proto/beans/GetTablesResp.java      |  221 ----
 .../drill/exec/proto/beans/GroupBySupport.java     |   53 -
 .../drill/exec/proto/beans/HandshakeStatus.java    |   55 -
 .../drill/exec/proto/beans/IdentifierCasing.java   |   55 -
 .../exec/proto/beans/InitializeFragments.java      |  175 ---
 .../org/apache/drill/exec/proto/beans/Jar.java     |  195 ---
 .../apache/drill/exec/proto/beans/LikeFilter.java  |  185 ---
 .../exec/proto/beans/MajorFragmentProfile.java     |  197 ---
 .../apache/drill/exec/proto/beans/MetricValue.java |  207 ---
 .../exec/proto/beans/MinorFragmentProfile.java     |  399 ------
 .../apache/drill/exec/proto/beans/NamePart.java    |  237 ----
 .../apache/drill/exec/proto/beans/NodeStatus.java  |  185 ---
 .../drill/exec/proto/beans/NullCollation.java      |   55 -
 .../drill/exec/proto/beans/OperatorProfile.java    |  339 -----
 .../drill/exec/proto/beans/OrderBySupport.java     |   51 -
 .../drill/exec/proto/beans/OuterJoinSupport.java   |   61 -
 .../drill/exec/proto/beans/ParsingError.java       |  229 ----
 .../drill/exec/proto/beans/PlanFragment.java       |  495 --------
 .../drill/exec/proto/beans/PreparedStatement.java  |  199 ---
 .../exec/proto/beans/PreparedStatementHandle.java  |  164 ---
 .../apache/drill/exec/proto/beans/Property.java    |  199 ---
 .../exec/proto/beans/QueryContextInformation.java  |  229 ----
 .../apache/drill/exec/proto/beans/QueryData.java   |  233 ----
 .../org/apache/drill/exec/proto/beans/QueryId.java |  185 ---
 .../apache/drill/exec/proto/beans/QueryInfo.java   |  322 -----
 .../drill/exec/proto/beans/QueryPlanFragments.java |  255 ----
 .../drill/exec/proto/beans/QueryProfile.java       |  666 ----------
 .../apache/drill/exec/proto/beans/QueryResult.java |  263 ----
 .../drill/exec/proto/beans/QueryResultsMode.java   |   47 -
 .../apache/drill/exec/proto/beans/QueryType.java   |   55 -
 .../drill/exec/proto/beans/RecordBatchDef.java     |  241 ----
 .../apache/drill/exec/proto/beans/Registry.java    |  175 ---
 .../drill/exec/proto/beans/RequestResults.java     |  187 ---
 .../drill/exec/proto/beans/RequestStatus.java      |   53 -
 .../exec/proto/beans/ResultColumnMetadata.java     |  559 --------
 .../org/apache/drill/exec/proto/beans/Roles.java   |  256 ----
 .../apache/drill/exec/proto/beans/RpcChannel.java  |   51 -
 .../drill/exec/proto/beans/RpcEndpointInfos.java   |  317 -----
 .../apache/drill/exec/proto/beans/RpcHeader.java   |  207 ---
 .../org/apache/drill/exec/proto/beans/RpcMode.java |   55 -
 .../org/apache/drill/exec/proto/beans/RpcType.java |   95 --
 .../apache/drill/exec/proto/beans/RunQuery.java    |  287 -----
 .../drill/exec/proto/beans/RuntimeFilterBDef.java  |  337 -----
 .../apache/drill/exec/proto/beans/SaslMessage.java |  208 ---
 .../apache/drill/exec/proto/beans/SaslStatus.java  |   55 -
 .../apache/drill/exec/proto/beans/SaslSupport.java |   51 -
 .../drill/exec/proto/beans/SchemaMetadata.java     |  251 ----
 .../drill/exec/proto/beans/SerializedField.java    |  289 -----
 .../apache/drill/exec/proto/beans/ServerMeta.java  | 1341 --------------------
 .../proto/beans/ServerPreparedStatementState.java  |  163 ---
 .../exec/proto/beans/StackTraceElementWrapper.java |  251 ----
 .../drill/exec/proto/beans/StreamProfile.java      |  207 ---
 .../drill/exec/proto/beans/SubQuerySupport.java    |   57 -
 .../drill/exec/proto/beans/TableMetadata.java      |  229 ----
 .../drill/exec/proto/beans/UnionSupport.java       |   51 -
 .../drill/exec/proto/beans/UserCredentials.java    |  163 ---
 .../drill/exec/proto/beans/UserProperties.java     |  175 ---
 .../drill/exec/proto/beans/UserToBitHandshake.java |  347 -----
 .../apache/drill/exec/proto/beans/ValueMode.java   |   51 -
 .../drill/exec/proto/beans/WorkQueueStatus.java    |  209 ---
 160 files changed, 994 insertions(+), 26063 deletions(-)
 delete mode 100644 exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/CompliantTextRecordReader.java
 delete mode 100644 exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/FieldVarCharOutput.java
 delete mode 100644 exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/HeaderBuilder.java
 delete mode 100644 exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/RepeatedVarCharOutput.java
 delete mode 100644 exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/StreamFinishedPseudoException.java
 delete mode 100644 exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextInput.java
 delete mode 100644 exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextOutput.java
 delete mode 100644 exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextParsingContext.java
 delete mode 100644 exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextParsingSettings.java
 delete mode 100644 exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextReader.java
 delete mode 100644 exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/package-info.java
 rename exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/{compliant/v3 => reader}/BaseFieldOutput.java (98%)
 rename exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/{compliant/v3 => reader}/CompliantTextBatchReader.java (98%)
 rename exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/{compliant/v3 => reader}/FieldVarCharOutput.java (97%)
 rename exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/{compliant/v3 => reader}/HeaderBuilder.java (99%)
 rename exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/{compliant/v3 => reader}/RepeatedVarCharOutput.java (98%)
 rename exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/{compliant/v3 => reader}/StreamFinishedPseudoException.java (83%)
 rename exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/{compliant/v3 => reader}/TextInput.java (97%)
 rename exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/{compliant/v3 => reader}/TextOutput.java (97%)
 rename exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/{compliant/v3 => reader}/TextParsingContext.java (97%)
 rename exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/{compliant/v3/TextParsingSettingsV3.java => reader/TextParsingSettings.java} (97%)
 rename exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/{compliant/v3 => reader}/TextReader.java (95%)
 rename exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/{compliant/v3 => reader}/package-info.java (93%)
 rename exec/java-exec/src/main/java/org/apache/drill/exec/store/{text/DrillTextRecordWriter.java => easy/text/writer/TextRecordWriter.java} (94%)
 delete mode 100644 exec/java-exec/src/main/java/org/apache/drill/exec/store/text/DrillTextRecordReader.java
 rename exec/java-exec/src/test/java/org/apache/drill/exec/store/{text => easy/text/compliant}/TestCsvHeader.java (99%)
 rename exec/java-exec/src/test/java/org/apache/drill/exec/store/{text => easy/text/compliant}/TestTextColumn.java (91%)
 rename exec/java-exec/src/test/java/org/apache/drill/exec/store/{text/TestNewTextReader.java => easy/text/compliant/TestTextReader.java} (91%)
 rename exec/java-exec/src/test/java/org/apache/drill/exec/store/{text => easy/text/compliant}/TextRecordReaderTest.java (97%)
 delete mode 100644 protocol/src/main/java/org/apache/drill/common/types/DataMode.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/common/types/MajorType.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/common/types/MinorType.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/Ack.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/BitClientHandshake.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/BitControlHandshake.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/BitServerHandshake.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/BitStatus.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/BitToUserHandshake.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/CatalogMetadata.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/CollateSupport.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/Collector.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/ColumnMetadata.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/ColumnSearchability.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/ColumnUpdatability.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/CompleteRpcMessage.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/ConvertSupport.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/CoreOperatorType.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/CorrelationNamesSupport.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/CreatePreparedStatementReq.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/CreatePreparedStatementResp.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/CustomMessage.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/DateTimeLiteralsSupport.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/DrillPBError.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/DrillServiceInstance.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/DrillbitEndpoint.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/ExceptionWrapper.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/FinishedReceiver.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/FragmentHandle.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/FragmentRecordBatch.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/FragmentState.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/FragmentStatus.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/GetCatalogsReq.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/GetCatalogsResp.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/GetColumnsReq.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/GetColumnsResp.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/GetQueryPlanFragments.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/GetSchemasReq.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/GetSchemasResp.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/GetServerMetaResp.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/GetTablesReq.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/GetTablesResp.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/GroupBySupport.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/HandshakeStatus.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/IdentifierCasing.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/InitializeFragments.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/Jar.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/LikeFilter.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/MajorFragmentProfile.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/MetricValue.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/MinorFragmentProfile.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/NamePart.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/NodeStatus.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/NullCollation.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/OperatorProfile.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/OrderBySupport.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/OuterJoinSupport.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/ParsingError.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/PlanFragment.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/PreparedStatement.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/PreparedStatementHandle.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/Property.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/QueryContextInformation.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/QueryData.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/QueryId.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/QueryInfo.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/QueryPlanFragments.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/QueryProfile.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/QueryResult.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/QueryResultsMode.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/QueryType.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/RecordBatchDef.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/Registry.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/RequestResults.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/RequestStatus.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/ResultColumnMetadata.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/Roles.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/RpcChannel.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/RpcEndpointInfos.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/RpcHeader.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/RpcMode.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/RpcType.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/RunQuery.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/RuntimeFilterBDef.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/SaslMessage.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/SaslStatus.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/SaslSupport.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/SchemaMetadata.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/SerializedField.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/ServerMeta.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/ServerPreparedStatementState.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/StackTraceElementWrapper.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/StreamProfile.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/SubQuerySupport.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/TableMetadata.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/UnionSupport.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/UserCredentials.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/UserProperties.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/UserToBitHandshake.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/ValueMode.java
 delete mode 100644 protocol/src/main/java/org/apache/drill/exec/proto/beans/WorkQueueStatus.java


[drill] 02/03: DRILL-7294: Regenerate protobufs

Posted by vo...@apache.org.
This is an automated email from the ASF dual-hosted git repository.

volodymyr pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/drill.git

commit 24ec98043ecdce6af8ccb054cf44ae943cd6cb4e
Author: Volodymyr Vysotskyi <vv...@gmail.com>
AuthorDate: Thu Jun 13 17:03:09 2019 +0300

    DRILL-7294: Regenerate protobufs
---
 .../org/apache/drill/common/types/DataMode.java    |   51 -
 .../org/apache/drill/common/types/MajorType.java   |  305 -----
 .../org/apache/drill/common/types/MinorType.java   |  123 --
 .../drill/common/types/SchemaTypeProtos.java       |   12 +-
 .../apache/drill/exec/proto/SchemaBitControl.java  |  120 +-
 .../org/apache/drill/exec/proto/SchemaBitData.java |   48 +-
 .../drill/exec/proto/SchemaCoordinationProtos.java |   36 +-
 .../apache/drill/exec/proto/SchemaExecProtos.java  |   24 +-
 .../drill/exec/proto/SchemaGeneralRPCProtos.java   |   36 +-
 .../drill/exec/proto/SchemaUserBitShared.java      |  264 ++--
 .../apache/drill/exec/proto/SchemaUserProtos.java  |  372 +++---
 .../org/apache/drill/exec/proto/beans/Ack.java     |  163 ---
 .../drill/exec/proto/beans/BitClientHandshake.java |  185 ---
 .../exec/proto/beans/BitControlHandshake.java      |  241 ----
 .../drill/exec/proto/beans/BitServerHandshake.java |  195 ---
 .../apache/drill/exec/proto/beans/BitStatus.java   |  175 ---
 .../drill/exec/proto/beans/BitToUserHandshake.java |  359 ------
 .../drill/exec/proto/beans/CatalogMetadata.java    |  207 ---
 .../drill/exec/proto/beans/CollateSupport.java     |   49 -
 .../apache/drill/exec/proto/beans/Collector.java   |  239 ----
 .../drill/exec/proto/beans/ColumnMetadata.java     |  515 --------
 .../exec/proto/beans/ColumnSearchability.java      |   55 -
 .../drill/exec/proto/beans/ColumnUpdatability.java |   51 -
 .../drill/exec/proto/beans/CompleteRpcMessage.java |  210 ---
 .../drill/exec/proto/beans/ConvertSupport.java     |  199 ---
 .../drill/exec/proto/beans/CoreOperatorType.java   |  171 ---
 .../exec/proto/beans/CorrelationNamesSupport.java  |   51 -
 .../proto/beans/CreatePreparedStatementReq.java    |  163 ---
 .../proto/beans/CreatePreparedStatementResp.java   |  211 ---
 .../drill/exec/proto/beans/CustomMessage.java      |  186 ---
 .../exec/proto/beans/DateTimeLiteralsSupport.java  |   79 --
 .../drill/exec/proto/beans/DrillPBError.java       |  341 -----
 .../exec/proto/beans/DrillServiceInstance.java     |  209 ---
 .../drill/exec/proto/beans/DrillbitEndpoint.java   |  351 -----
 .../drill/exec/proto/beans/ExceptionWrapper.java   |  243 ----
 .../drill/exec/proto/beans/FinishedReceiver.java   |  189 ---
 .../drill/exec/proto/beans/FragmentHandle.java     |  233 ----
 .../exec/proto/beans/FragmentRecordBatch.java      |  309 -----
 .../drill/exec/proto/beans/FragmentState.java      |   59 -
 .../drill/exec/proto/beans/FragmentStatus.java     |  189 ---
 .../drill/exec/proto/beans/GetCatalogsReq.java     |  165 ---
 .../drill/exec/proto/beans/GetCatalogsResp.java    |  221 ----
 .../drill/exec/proto/beans/GetColumnsReq.java      |  237 ----
 .../drill/exec/proto/beans/GetColumnsResp.java     |  221 ----
 .../exec/proto/beans/GetQueryPlanFragments.java    |  218 ----
 .../drill/exec/proto/beans/GetSchemasReq.java      |  189 ---
 .../drill/exec/proto/beans/GetSchemasResp.java     |  221 ----
 .../drill/exec/proto/beans/GetServerMetaResp.java  |  211 ---
 .../drill/exec/proto/beans/GetTablesReq.java       |  245 ----
 .../drill/exec/proto/beans/GetTablesResp.java      |  221 ----
 .../drill/exec/proto/beans/GroupBySupport.java     |   53 -
 .../drill/exec/proto/beans/HandshakeStatus.java    |   55 -
 .../drill/exec/proto/beans/IdentifierCasing.java   |   55 -
 .../exec/proto/beans/InitializeFragments.java      |  175 ---
 .../org/apache/drill/exec/proto/beans/Jar.java     |  195 ---
 .../apache/drill/exec/proto/beans/LikeFilter.java  |  185 ---
 .../exec/proto/beans/MajorFragmentProfile.java     |  197 ---
 .../apache/drill/exec/proto/beans/MetricValue.java |  207 ---
 .../exec/proto/beans/MinorFragmentProfile.java     |  399 ------
 .../apache/drill/exec/proto/beans/NamePart.java    |  237 ----
 .../apache/drill/exec/proto/beans/NodeStatus.java  |  185 ---
 .../drill/exec/proto/beans/NullCollation.java      |   55 -
 .../drill/exec/proto/beans/OperatorProfile.java    |  339 -----
 .../drill/exec/proto/beans/OrderBySupport.java     |   51 -
 .../drill/exec/proto/beans/OuterJoinSupport.java   |   61 -
 .../drill/exec/proto/beans/ParsingError.java       |  229 ----
 .../drill/exec/proto/beans/PlanFragment.java       |  495 --------
 .../drill/exec/proto/beans/PreparedStatement.java  |  199 ---
 .../exec/proto/beans/PreparedStatementHandle.java  |  164 ---
 .../apache/drill/exec/proto/beans/Property.java    |  199 ---
 .../exec/proto/beans/QueryContextInformation.java  |  229 ----
 .../apache/drill/exec/proto/beans/QueryData.java   |  233 ----
 .../org/apache/drill/exec/proto/beans/QueryId.java |  185 ---
 .../apache/drill/exec/proto/beans/QueryInfo.java   |  322 -----
 .../drill/exec/proto/beans/QueryPlanFragments.java |  255 ----
 .../drill/exec/proto/beans/QueryProfile.java       |  666 ----------
 .../apache/drill/exec/proto/beans/QueryResult.java |  263 ----
 .../drill/exec/proto/beans/QueryResultsMode.java   |   47 -
 .../apache/drill/exec/proto/beans/QueryType.java   |   55 -
 .../drill/exec/proto/beans/RecordBatchDef.java     |  241 ----
 .../apache/drill/exec/proto/beans/Registry.java    |  175 ---
 .../drill/exec/proto/beans/RequestResults.java     |  187 ---
 .../drill/exec/proto/beans/RequestStatus.java      |   53 -
 .../exec/proto/beans/ResultColumnMetadata.java     |  559 --------
 .../org/apache/drill/exec/proto/beans/Roles.java   |  256 ----
 .../apache/drill/exec/proto/beans/RpcChannel.java  |   51 -
 .../drill/exec/proto/beans/RpcEndpointInfos.java   |  317 -----
 .../apache/drill/exec/proto/beans/RpcHeader.java   |  207 ---
 .../org/apache/drill/exec/proto/beans/RpcMode.java |   55 -
 .../org/apache/drill/exec/proto/beans/RpcType.java |   95 --
 .../apache/drill/exec/proto/beans/RunQuery.java    |  287 -----
 .../drill/exec/proto/beans/RuntimeFilterBDef.java  |  337 -----
 .../apache/drill/exec/proto/beans/SaslMessage.java |  208 ---
 .../apache/drill/exec/proto/beans/SaslStatus.java  |   55 -
 .../apache/drill/exec/proto/beans/SaslSupport.java |   51 -
 .../drill/exec/proto/beans/SchemaMetadata.java     |  251 ----
 .../drill/exec/proto/beans/SerializedField.java    |  289 -----
 .../apache/drill/exec/proto/beans/ServerMeta.java  | 1341 --------------------
 .../proto/beans/ServerPreparedStatementState.java  |  163 ---
 .../exec/proto/beans/StackTraceElementWrapper.java |  251 ----
 .../drill/exec/proto/beans/StreamProfile.java      |  207 ---
 .../drill/exec/proto/beans/SubQuerySupport.java    |   57 -
 .../drill/exec/proto/beans/TableMetadata.java      |  229 ----
 .../drill/exec/proto/beans/UnionSupport.java       |   51 -
 .../drill/exec/proto/beans/UserCredentials.java    |  163 ---
 .../drill/exec/proto/beans/UserProperties.java     |  175 ---
 .../drill/exec/proto/beans/UserToBitHandshake.java |  347 -----
 .../apache/drill/exec/proto/beans/ValueMode.java   |   51 -
 .../drill/exec/proto/beans/WorkQueueStatus.java    |  209 ---
 109 files changed, 456 insertions(+), 21819 deletions(-)

diff --git a/protocol/src/main/java/org/apache/drill/common/types/DataMode.java b/protocol/src/main/java/org/apache/drill/common/types/DataMode.java
deleted file mode 100644
index f1187d3..0000000
--- a/protocol/src/main/java/org/apache/drill/common/types/DataMode.java
+++ /dev/null
@@ -1,51 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.common.types;
-
-public enum DataMode implements com.dyuproject.protostuff.EnumLite<DataMode>
-{
-    OPTIONAL(0),
-    REQUIRED(1),
-    REPEATED(2);
-    
-    public final int number;
-    
-    private DataMode (int number)
-    {
-        this.number = number;
-    }
-    
-    public int getNumber()
-    {
-        return number;
-    }
-    
-    public static DataMode valueOf(int number)
-    {
-        switch(number) 
-        {
-            case 0: return OPTIONAL;
-            case 1: return REQUIRED;
-            case 2: return REPEATED;
-            default: return null;
-        }
-    }
-}
diff --git a/protocol/src/main/java/org/apache/drill/common/types/MajorType.java b/protocol/src/main/java/org/apache/drill/common/types/MajorType.java
deleted file mode 100644
index 104af34..0000000
--- a/protocol/src/main/java/org/apache/drill/common/types/MajorType.java
+++ /dev/null
@@ -1,305 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.common.types;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-import java.util.ArrayList;
-import java.util.List;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class MajorType implements Externalizable, Message<MajorType>, Schema<MajorType>
-{
-
-    public static Schema<MajorType> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static MajorType getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final MajorType DEFAULT_INSTANCE = new MajorType();
-
-    
-    private MinorType minorType;
-    private DataMode mode;
-    private int width;
-    private int precision;
-    private int scale;
-    private int timeZone;
-    private List<MinorType> subType;
-
-    public MajorType()
-    {
-        
-    }
-
-    // getters and setters
-
-    // minorType
-
-    public MinorType getMinorType()
-    {
-        return minorType == null ? MinorType.LATE : minorType;
-    }
-
-    public MajorType setMinorType(MinorType minorType)
-    {
-        this.minorType = minorType;
-        return this;
-    }
-
-    // mode
-
-    public DataMode getMode()
-    {
-        return mode == null ? DataMode.OPTIONAL : mode;
-    }
-
-    public MajorType setMode(DataMode mode)
-    {
-        this.mode = mode;
-        return this;
-    }
-
-    // width
-
-    public int getWidth()
-    {
-        return width;
-    }
-
-    public MajorType setWidth(int width)
-    {
-        this.width = width;
-        return this;
-    }
-
-    // precision
-
-    public int getPrecision()
-    {
-        return precision;
-    }
-
-    public MajorType setPrecision(int precision)
-    {
-        this.precision = precision;
-        return this;
-    }
-
-    // scale
-
-    public int getScale()
-    {
-        return scale;
-    }
-
-    public MajorType setScale(int scale)
-    {
-        this.scale = scale;
-        return this;
-    }
-
-    // timeZone
-
-    public int getTimeZone()
-    {
-        return timeZone;
-    }
-
-    public MajorType setTimeZone(int timeZone)
-    {
-        this.timeZone = timeZone;
-        return this;
-    }
-
-    // subType
-
-    public List<MinorType> getSubTypeList()
-    {
-        return subType;
-    }
-
-    public MajorType setSubTypeList(List<MinorType> subType)
-    {
-        this.subType = subType;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<MajorType> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public MajorType newMessage()
-    {
-        return new MajorType();
-    }
-
-    public Class<MajorType> typeClass()
-    {
-        return MajorType.class;
-    }
-
-    public String messageName()
-    {
-        return MajorType.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return MajorType.class.getName();
-    }
-
-    public boolean isInitialized(MajorType message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, MajorType message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.minorType = MinorType.valueOf(input.readEnum());
-                    break;
-                case 2:
-                    message.mode = DataMode.valueOf(input.readEnum());
-                    break;
-                case 3:
-                    message.width = input.readInt32();
-                    break;
-                case 4:
-                    message.precision = input.readInt32();
-                    break;
-                case 5:
-                    message.scale = input.readInt32();
-                    break;
-                case 6:
-                    message.timeZone = input.readInt32();
-                    break;
-                case 7:
-                    if(message.subType == null)
-                        message.subType = new ArrayList<MinorType>();
-                    message.subType.add(MinorType.valueOf(input.readEnum()));
-                    break;
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, MajorType message) throws IOException
-    {
-        if(message.minorType != null)
-             output.writeEnum(1, message.minorType.number, false);
-
-        if(message.mode != null)
-             output.writeEnum(2, message.mode.number, false);
-
-        if(message.width != 0)
-            output.writeInt32(3, message.width, false);
-
-        if(message.precision != 0)
-            output.writeInt32(4, message.precision, false);
-
-        if(message.scale != 0)
-            output.writeInt32(5, message.scale, false);
-
-        if(message.timeZone != 0)
-            output.writeInt32(6, message.timeZone, false);
-
-        if(message.subType != null)
-        {
-            for(MinorType subType : message.subType)
-            {
-                if(subType != null)
-                    output.writeEnum(7, subType.number, true);
-            }
-        }
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "minorType";
-            case 2: return "mode";
-            case 3: return "width";
-            case 4: return "precision";
-            case 5: return "scale";
-            case 6: return "timeZone";
-            case 7: return "subType";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("minorType", 1);
-        __fieldMap.put("mode", 2);
-        __fieldMap.put("width", 3);
-        __fieldMap.put("precision", 4);
-        __fieldMap.put("scale", 5);
-        __fieldMap.put("timeZone", 6);
-        __fieldMap.put("subType", 7);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/common/types/MinorType.java b/protocol/src/main/java/org/apache/drill/common/types/MinorType.java
deleted file mode 100644
index b26cb08..0000000
--- a/protocol/src/main/java/org/apache/drill/common/types/MinorType.java
+++ /dev/null
@@ -1,123 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.common.types;
-
-public enum MinorType implements com.dyuproject.protostuff.EnumLite<MinorType>
-{
-    LATE(0),
-    MAP(1),
-    TINYINT(3),
-    SMALLINT(4),
-    INT(5),
-    BIGINT(6),
-    DECIMAL9(7),
-    DECIMAL18(8),
-    DECIMAL28SPARSE(9),
-    DECIMAL38SPARSE(10),
-    MONEY(11),
-    DATE(12),
-    TIME(13),
-    TIMETZ(14),
-    TIMESTAMPTZ(15),
-    TIMESTAMP(16),
-    INTERVAL(17),
-    FLOAT4(18),
-    FLOAT8(19),
-    BIT(20),
-    FIXEDCHAR(21),
-    FIXED16CHAR(22),
-    FIXEDBINARY(23),
-    VARCHAR(24),
-    VAR16CHAR(25),
-    VARBINARY(26),
-    UINT1(29),
-    UINT2(30),
-    UINT4(31),
-    UINT8(32),
-    DECIMAL28DENSE(33),
-    DECIMAL38DENSE(34),
-    NULL(37),
-    INTERVALYEAR(38),
-    INTERVALDAY(39),
-    LIST(40),
-    GENERIC_OBJECT(41),
-    UNION(42),
-    VARDECIMAL(43);
-    
-    public final int number;
-    
-    private MinorType (int number)
-    {
-        this.number = number;
-    }
-    
-    public int getNumber()
-    {
-        return number;
-    }
-    
-    public static MinorType valueOf(int number)
-    {
-        switch(number) 
-        {
-            case 0: return LATE;
-            case 1: return MAP;
-            case 3: return TINYINT;
-            case 4: return SMALLINT;
-            case 5: return INT;
-            case 6: return BIGINT;
-            case 7: return DECIMAL9;
-            case 8: return DECIMAL18;
-            case 9: return DECIMAL28SPARSE;
-            case 10: return DECIMAL38SPARSE;
-            case 11: return MONEY;
-            case 12: return DATE;
-            case 13: return TIME;
-            case 14: return TIMETZ;
-            case 15: return TIMESTAMPTZ;
-            case 16: return TIMESTAMP;
-            case 17: return INTERVAL;
-            case 18: return FLOAT4;
-            case 19: return FLOAT8;
-            case 20: return BIT;
-            case 21: return FIXEDCHAR;
-            case 22: return FIXED16CHAR;
-            case 23: return FIXEDBINARY;
-            case 24: return VARCHAR;
-            case 25: return VAR16CHAR;
-            case 26: return VARBINARY;
-            case 29: return UINT1;
-            case 30: return UINT2;
-            case 31: return UINT4;
-            case 32: return UINT8;
-            case 33: return DECIMAL28DENSE;
-            case 34: return DECIMAL38DENSE;
-            case 37: return NULL;
-            case 38: return INTERVALYEAR;
-            case 39: return INTERVALDAY;
-            case 40: return LIST;
-            case 41: return GENERIC_OBJECT;
-            case 42: return UNION;
-            case 43: return VARDECIMAL;
-            default: return null;
-        }
-    }
-}
diff --git a/protocol/src/main/java/org/apache/drill/common/types/SchemaTypeProtos.java b/protocol/src/main/java/org/apache/drill/common/types/SchemaTypeProtos.java
index ae30e40..71e6e51 100644
--- a/protocol/src/main/java/org/apache/drill/common/types/SchemaTypeProtos.java
+++ b/protocol/src/main/java/org/apache/drill/common/types/SchemaTypeProtos.java
@@ -31,9 +31,9 @@ public final class SchemaTypeProtos
         public static final org.apache.drill.common.types.SchemaTypeProtos.MajorType.BuilderSchema MERGE =
             new org.apache.drill.common.types.SchemaTypeProtos.MajorType.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.common.types.TypeProtos.MajorType>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.common.types.TypeProtos.MajorType>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.common.types.TypeProtos.MajorType message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.common.types.TypeProtos.MajorType message) throws java.io.IOException
             {
                 if(message.hasMinorType())
                     output.writeEnum(1, message.getMinorType().getNumber(), false);
@@ -75,12 +75,12 @@ public final class SchemaTypeProtos
                 return org.apache.drill.common.types.TypeProtos.MajorType.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.common.types.TypeProtos.MajorType message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.common.types.TypeProtos.MajorType message) throws java.io.IOException {}
             public org.apache.drill.common.types.TypeProtos.MajorType newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.common.types.TypeProtos.MajorType.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.common.types.TypeProtos.MajorType.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.common.types.TypeProtos.MajorType.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.common.types.TypeProtos.MajorType.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -143,7 +143,7 @@ public final class SchemaTypeProtos
                 return org.apache.drill.common.types.TypeProtos.MajorType.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.common.types.TypeProtos.MajorType.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.common.types.TypeProtos.MajorType.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/SchemaBitControl.java b/protocol/src/main/java/org/apache/drill/exec/proto/SchemaBitControl.java
index 4754195..23b2638 100644
--- a/protocol/src/main/java/org/apache/drill/exec/proto/SchemaBitControl.java
+++ b/protocol/src/main/java/org/apache/drill/exec/proto/SchemaBitControl.java
@@ -31,9 +31,9 @@ public final class SchemaBitControl
         public static final org.apache.drill.exec.proto.SchemaBitControl.BitControlHandshake.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaBitControl.BitControlHandshake.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitControl.BitControlHandshake>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitControl.BitControlHandshake>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitControl.BitControlHandshake message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitControl.BitControlHandshake message) throws java.io.IOException
             {
                 if(message.hasRpcVersion())
                     output.writeInt32(1, message.getRpcVersion(), false);
@@ -70,12 +70,12 @@ public final class SchemaBitControl
                 return org.apache.drill.exec.proto.BitControl.BitControlHandshake.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitControl.BitControlHandshake message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitControl.BitControlHandshake message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.BitControl.BitControlHandshake newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitControl.BitControlHandshake.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitControl.BitControlHandshake.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitControl.BitControlHandshake.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitControl.BitControlHandshake.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -130,7 +130,7 @@ public final class SchemaBitControl
                 return org.apache.drill.exec.proto.BitControl.BitControlHandshake.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitControl.BitControlHandshake.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitControl.BitControlHandshake.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -165,9 +165,9 @@ public final class SchemaBitControl
         public static final org.apache.drill.exec.proto.SchemaBitControl.BitStatus.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaBitControl.BitStatus.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitControl.BitStatus>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitControl.BitStatus>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitControl.BitStatus message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitControl.BitStatus message) throws java.io.IOException
             {
                 for(org.apache.drill.exec.proto.BitControl.FragmentStatus fragmentStatus : message.getFragmentStatusList())
                     output.writeObject(1, fragmentStatus, org.apache.drill.exec.proto.SchemaBitControl.FragmentStatus.WRITE, true);
@@ -198,12 +198,12 @@ public final class SchemaBitControl
                 return org.apache.drill.exec.proto.BitControl.BitStatus.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitControl.BitStatus message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitControl.BitStatus message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.BitControl.BitStatus newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitControl.BitStatus.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitControl.BitStatus.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitControl.BitStatus.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitControl.BitStatus.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -249,7 +249,7 @@ public final class SchemaBitControl
                 return org.apache.drill.exec.proto.BitControl.BitStatus.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitControl.BitStatus.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitControl.BitStatus.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -278,9 +278,9 @@ public final class SchemaBitControl
         public static final org.apache.drill.exec.proto.SchemaBitControl.FragmentStatus.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaBitControl.FragmentStatus.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitControl.FragmentStatus>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitControl.FragmentStatus>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitControl.FragmentStatus message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitControl.FragmentStatus message) throws java.io.IOException
             {
                 if(message.hasProfile())
                     output.writeObject(1, message.getProfile(), org.apache.drill.exec.proto.SchemaUserBitShared.MinorFragmentProfile.WRITE, false);
@@ -314,12 +314,12 @@ public final class SchemaBitControl
                 return org.apache.drill.exec.proto.BitControl.FragmentStatus.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitControl.FragmentStatus message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitControl.FragmentStatus message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.BitControl.FragmentStatus newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitControl.FragmentStatus.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitControl.FragmentStatus.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitControl.FragmentStatus.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitControl.FragmentStatus.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -369,7 +369,7 @@ public final class SchemaBitControl
                 return org.apache.drill.exec.proto.BitControl.FragmentStatus.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitControl.FragmentStatus.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitControl.FragmentStatus.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -400,9 +400,9 @@ public final class SchemaBitControl
         public static final org.apache.drill.exec.proto.SchemaBitControl.InitializeFragments.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaBitControl.InitializeFragments.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitControl.InitializeFragments>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitControl.InitializeFragments>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitControl.InitializeFragments message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitControl.InitializeFragments message) throws java.io.IOException
             {
                 for(org.apache.drill.exec.proto.BitControl.PlanFragment fragment : message.getFragmentList())
                     output.writeObject(1, fragment, org.apache.drill.exec.proto.SchemaBitControl.PlanFragment.WRITE, true);
@@ -433,12 +433,12 @@ public final class SchemaBitControl
                 return org.apache.drill.exec.proto.BitControl.InitializeFragments.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitControl.InitializeFragments message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitControl.InitializeFragments message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.BitControl.InitializeFragments newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitControl.InitializeFragments.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitControl.InitializeFragments.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitControl.InitializeFragments.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitControl.InitializeFragments.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -484,7 +484,7 @@ public final class SchemaBitControl
                 return org.apache.drill.exec.proto.BitControl.InitializeFragments.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitControl.InitializeFragments.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitControl.InitializeFragments.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -513,9 +513,9 @@ public final class SchemaBitControl
         public static final org.apache.drill.exec.proto.SchemaBitControl.CustomMessage.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaBitControl.CustomMessage.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitControl.CustomMessage>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitControl.CustomMessage>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitControl.CustomMessage message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitControl.CustomMessage message) throws java.io.IOException
             {
                 if(message.hasType())
                     output.writeInt32(1, message.getType(), false);
@@ -548,12 +548,12 @@ public final class SchemaBitControl
                 return org.apache.drill.exec.proto.BitControl.CustomMessage.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitControl.CustomMessage message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitControl.CustomMessage message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.BitControl.CustomMessage newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitControl.CustomMessage.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitControl.CustomMessage.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitControl.CustomMessage.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitControl.CustomMessage.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -601,7 +601,7 @@ public final class SchemaBitControl
                 return org.apache.drill.exec.proto.BitControl.CustomMessage.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitControl.CustomMessage.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitControl.CustomMessage.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -632,9 +632,9 @@ public final class SchemaBitControl
         public static final org.apache.drill.exec.proto.SchemaBitControl.PlanFragment.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaBitControl.PlanFragment.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitControl.PlanFragment>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitControl.PlanFragment>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitControl.PlanFragment message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitControl.PlanFragment message) throws java.io.IOException
             {
                 if(message.hasHandle())
                     output.writeObject(1, message.getHandle(), org.apache.drill.exec.proto.SchemaExecProtos.FragmentHandle.WRITE, false);
@@ -698,12 +698,12 @@ public final class SchemaBitControl
                 return org.apache.drill.exec.proto.BitControl.PlanFragment.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitControl.PlanFragment message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitControl.PlanFragment message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.BitControl.PlanFragment newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitControl.PlanFragment.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitControl.PlanFragment.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitControl.PlanFragment.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitControl.PlanFragment.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -796,7 +796,7 @@ public final class SchemaBitControl
                 return org.apache.drill.exec.proto.BitControl.PlanFragment.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitControl.PlanFragment.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitControl.PlanFragment.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -853,9 +853,9 @@ public final class SchemaBitControl
         public static final org.apache.drill.exec.proto.SchemaBitControl.Collector.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaBitControl.Collector.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitControl.Collector>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitControl.Collector>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitControl.Collector message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitControl.Collector message) throws java.io.IOException
             {
                 if(message.hasOppositeMajorFragmentId())
                     output.writeInt32(1, message.getOppositeMajorFragmentId(), false);
@@ -891,12 +891,12 @@ public final class SchemaBitControl
                 return org.apache.drill.exec.proto.BitControl.Collector.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitControl.Collector message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitControl.Collector message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.BitControl.Collector newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitControl.Collector.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitControl.Collector.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitControl.Collector.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitControl.Collector.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -950,7 +950,7 @@ public final class SchemaBitControl
                 return org.apache.drill.exec.proto.BitControl.Collector.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitControl.Collector.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitControl.Collector.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -985,9 +985,9 @@ public final class SchemaBitControl
         public static final org.apache.drill.exec.proto.SchemaBitControl.QueryContextInformation.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaBitControl.QueryContextInformation.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitControl.QueryContextInformation>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitControl.QueryContextInformation>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitControl.QueryContextInformation message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitControl.QueryContextInformation message) throws java.io.IOException
             {
                 if(message.hasQueryStartTime())
                     output.writeInt64(1, message.getQueryStartTime(), false);
@@ -1023,12 +1023,12 @@ public final class SchemaBitControl
                 return org.apache.drill.exec.proto.BitControl.QueryContextInformation.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitControl.QueryContextInformation message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitControl.QueryContextInformation message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.BitControl.QueryContextInformation newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitControl.QueryContextInformation.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitControl.QueryContextInformation.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitControl.QueryContextInformation.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitControl.QueryContextInformation.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -1082,7 +1082,7 @@ public final class SchemaBitControl
                 return org.apache.drill.exec.proto.BitControl.QueryContextInformation.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitControl.QueryContextInformation.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitControl.QueryContextInformation.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -1117,9 +1117,9 @@ public final class SchemaBitControl
         public static final org.apache.drill.exec.proto.SchemaBitControl.WorkQueueStatus.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaBitControl.WorkQueueStatus.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitControl.WorkQueueStatus>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitControl.WorkQueueStatus>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitControl.WorkQueueStatus message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitControl.WorkQueueStatus message) throws java.io.IOException
             {
                 if(message.hasEndpoint())
                     output.writeObject(1, message.getEndpoint(), org.apache.drill.exec.proto.SchemaCoordinationProtos.DrillbitEndpoint.WRITE, false);
@@ -1154,12 +1154,12 @@ public final class SchemaBitControl
                 return org.apache.drill.exec.proto.BitControl.WorkQueueStatus.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitControl.WorkQueueStatus message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitControl.WorkQueueStatus message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.BitControl.WorkQueueStatus newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitControl.WorkQueueStatus.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitControl.WorkQueueStatus.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitControl.WorkQueueStatus.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitControl.WorkQueueStatus.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -1211,7 +1211,7 @@ public final class SchemaBitControl
                 return org.apache.drill.exec.proto.BitControl.WorkQueueStatus.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitControl.WorkQueueStatus.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitControl.WorkQueueStatus.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -1244,9 +1244,9 @@ public final class SchemaBitControl
         public static final org.apache.drill.exec.proto.SchemaBitControl.FinishedReceiver.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaBitControl.FinishedReceiver.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitControl.FinishedReceiver>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitControl.FinishedReceiver>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitControl.FinishedReceiver message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitControl.FinishedReceiver message) throws java.io.IOException
             {
                 if(message.hasReceiver())
                     output.writeObject(1, message.getReceiver(), org.apache.drill.exec.proto.SchemaExecProtos.FragmentHandle.WRITE, false);
@@ -1280,12 +1280,12 @@ public final class SchemaBitControl
                 return org.apache.drill.exec.proto.BitControl.FinishedReceiver.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitControl.FinishedReceiver message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitControl.FinishedReceiver message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.BitControl.FinishedReceiver newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitControl.FinishedReceiver.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitControl.FinishedReceiver.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitControl.FinishedReceiver.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitControl.FinishedReceiver.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -1335,7 +1335,7 @@ public final class SchemaBitControl
                 return org.apache.drill.exec.proto.BitControl.FinishedReceiver.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitControl.FinishedReceiver.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitControl.FinishedReceiver.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/SchemaBitData.java b/protocol/src/main/java/org/apache/drill/exec/proto/SchemaBitData.java
index ecf0f18..091c690 100644
--- a/protocol/src/main/java/org/apache/drill/exec/proto/SchemaBitData.java
+++ b/protocol/src/main/java/org/apache/drill/exec/proto/SchemaBitData.java
@@ -31,9 +31,9 @@ public final class SchemaBitData
         public static final org.apache.drill.exec.proto.SchemaBitData.BitClientHandshake.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaBitData.BitClientHandshake.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitData.BitClientHandshake>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitData.BitClientHandshake>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitData.BitClientHandshake message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitData.BitClientHandshake message) throws java.io.IOException
             {
                 if(message.hasRpcVersion())
                     output.writeInt32(1, message.getRpcVersion(), false);
@@ -65,12 +65,12 @@ public final class SchemaBitData
                 return org.apache.drill.exec.proto.BitData.BitClientHandshake.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitData.BitClientHandshake message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitData.BitClientHandshake message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.BitData.BitClientHandshake newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitData.BitClientHandshake.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitData.BitClientHandshake.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitData.BitClientHandshake.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitData.BitClientHandshake.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -118,7 +118,7 @@ public final class SchemaBitData
                 return org.apache.drill.exec.proto.BitData.BitClientHandshake.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitData.BitClientHandshake.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitData.BitClientHandshake.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -149,9 +149,9 @@ public final class SchemaBitData
         public static final org.apache.drill.exec.proto.SchemaBitData.BitServerHandshake.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaBitData.BitServerHandshake.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitData.BitServerHandshake>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitData.BitServerHandshake>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitData.BitServerHandshake message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitData.BitServerHandshake message) throws java.io.IOException
             {
                 if(message.hasRpcVersion())
                     output.writeInt32(1, message.getRpcVersion(), false);
@@ -183,12 +183,12 @@ public final class SchemaBitData
                 return org.apache.drill.exec.proto.BitData.BitServerHandshake.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitData.BitServerHandshake message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitData.BitServerHandshake message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.BitData.BitServerHandshake newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitData.BitServerHandshake.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitData.BitServerHandshake.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitData.BitServerHandshake.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitData.BitServerHandshake.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -236,7 +236,7 @@ public final class SchemaBitData
                 return org.apache.drill.exec.proto.BitData.BitServerHandshake.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitData.BitServerHandshake.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitData.BitServerHandshake.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -267,9 +267,9 @@ public final class SchemaBitData
         public static final org.apache.drill.exec.proto.SchemaBitData.FragmentRecordBatch.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaBitData.FragmentRecordBatch.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitData.FragmentRecordBatch>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitData.FragmentRecordBatch>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitData.FragmentRecordBatch message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitData.FragmentRecordBatch message) throws java.io.IOException
             {
                 if(message.hasQueryId())
                     output.writeObject(1, message.getQueryId(), org.apache.drill.exec.proto.SchemaUserBitShared.QueryId.WRITE, false);
@@ -313,12 +313,12 @@ public final class SchemaBitData
                 return org.apache.drill.exec.proto.BitData.FragmentRecordBatch.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitData.FragmentRecordBatch message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitData.FragmentRecordBatch message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.BitData.FragmentRecordBatch newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitData.FragmentRecordBatch.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitData.FragmentRecordBatch.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitData.FragmentRecordBatch.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitData.FragmentRecordBatch.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -383,7 +383,7 @@ public final class SchemaBitData
                 return org.apache.drill.exec.proto.BitData.FragmentRecordBatch.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitData.FragmentRecordBatch.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitData.FragmentRecordBatch.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -424,9 +424,9 @@ public final class SchemaBitData
         public static final org.apache.drill.exec.proto.SchemaBitData.RuntimeFilterBDef.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaBitData.RuntimeFilterBDef.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitData.RuntimeFilterBDef>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitData.RuntimeFilterBDef>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitData.RuntimeFilterBDef message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitData.RuntimeFilterBDef message) throws java.io.IOException
             {
                 if(message.hasQueryId())
                     output.writeObject(1, message.getQueryId(), org.apache.drill.exec.proto.SchemaUserBitShared.QueryId.WRITE, false);
@@ -471,12 +471,12 @@ public final class SchemaBitData
                 return org.apache.drill.exec.proto.BitData.RuntimeFilterBDef.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitData.RuntimeFilterBDef message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitData.RuntimeFilterBDef message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.BitData.RuntimeFilterBDef newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.BitData.RuntimeFilterBDef.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.BitData.RuntimeFilterBDef.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.BitData.RuntimeFilterBDef.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.BitData.RuntimeFilterBDef.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -543,7 +543,7 @@ public final class SchemaBitData
                 return org.apache.drill.exec.proto.BitData.RuntimeFilterBDef.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.BitData.RuntimeFilterBDef.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.BitData.RuntimeFilterBDef.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/SchemaCoordinationProtos.java b/protocol/src/main/java/org/apache/drill/exec/proto/SchemaCoordinationProtos.java
index eb29ecc..fbf1116 100644
--- a/protocol/src/main/java/org/apache/drill/exec/proto/SchemaCoordinationProtos.java
+++ b/protocol/src/main/java/org/apache/drill/exec/proto/SchemaCoordinationProtos.java
@@ -31,9 +31,9 @@ public final class SchemaCoordinationProtos
         public static final org.apache.drill.exec.proto.SchemaCoordinationProtos.DrillbitEndpoint.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaCoordinationProtos.DrillbitEndpoint.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.CoordinationProtos.DrillbitEndpoint>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.CoordinationProtos.DrillbitEndpoint>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.CoordinationProtos.DrillbitEndpoint message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.CoordinationProtos.DrillbitEndpoint message) throws java.io.IOException
             {
                 if(message.hasAddress())
                     output.writeString(1, message.getAddress(), false);
@@ -78,12 +78,12 @@ public final class SchemaCoordinationProtos
                 return org.apache.drill.exec.proto.CoordinationProtos.DrillbitEndpoint.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.CoordinationProtos.DrillbitEndpoint message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.CoordinationProtos.DrillbitEndpoint message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.CoordinationProtos.DrillbitEndpoint newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.CoordinationProtos.DrillbitEndpoint.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.CoordinationProtos.DrillbitEndpoint.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.CoordinationProtos.DrillbitEndpoint.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.CoordinationProtos.DrillbitEndpoint.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -150,7 +150,7 @@ public final class SchemaCoordinationProtos
                 return org.apache.drill.exec.proto.CoordinationProtos.DrillbitEndpoint.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.CoordinationProtos.DrillbitEndpoint.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.CoordinationProtos.DrillbitEndpoint.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -193,9 +193,9 @@ public final class SchemaCoordinationProtos
         public static final org.apache.drill.exec.proto.SchemaCoordinationProtos.DrillServiceInstance.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaCoordinationProtos.DrillServiceInstance.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.CoordinationProtos.DrillServiceInstance>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.CoordinationProtos.DrillServiceInstance>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.CoordinationProtos.DrillServiceInstance message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.CoordinationProtos.DrillServiceInstance message) throws java.io.IOException
             {
                 if(message.hasId())
                     output.writeString(1, message.getId(), false);
@@ -230,12 +230,12 @@ public final class SchemaCoordinationProtos
                 return org.apache.drill.exec.proto.CoordinationProtos.DrillServiceInstance.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.CoordinationProtos.DrillServiceInstance message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.CoordinationProtos.DrillServiceInstance message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.CoordinationProtos.DrillServiceInstance newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.CoordinationProtos.DrillServiceInstance.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.CoordinationProtos.DrillServiceInstance.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.CoordinationProtos.DrillServiceInstance.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.CoordinationProtos.DrillServiceInstance.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -287,7 +287,7 @@ public final class SchemaCoordinationProtos
                 return org.apache.drill.exec.proto.CoordinationProtos.DrillServiceInstance.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.CoordinationProtos.DrillServiceInstance.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.CoordinationProtos.DrillServiceInstance.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -320,9 +320,9 @@ public final class SchemaCoordinationProtos
         public static final org.apache.drill.exec.proto.SchemaCoordinationProtos.Roles.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaCoordinationProtos.Roles.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.CoordinationProtos.Roles>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.CoordinationProtos.Roles>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.CoordinationProtos.Roles message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.CoordinationProtos.Roles message) throws java.io.IOException
             {
                 if(message.hasSqlQuery())
                     output.writeBool(1, message.getSqlQuery(), false);
@@ -360,12 +360,12 @@ public final class SchemaCoordinationProtos
                 return org.apache.drill.exec.proto.CoordinationProtos.Roles.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.CoordinationProtos.Roles message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.CoordinationProtos.Roles message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.CoordinationProtos.Roles newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.CoordinationProtos.Roles.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.CoordinationProtos.Roles.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.CoordinationProtos.Roles.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.CoordinationProtos.Roles.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -422,7 +422,7 @@ public final class SchemaCoordinationProtos
                 return org.apache.drill.exec.proto.CoordinationProtos.Roles.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.CoordinationProtos.Roles.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.CoordinationProtos.Roles.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/SchemaExecProtos.java b/protocol/src/main/java/org/apache/drill/exec/proto/SchemaExecProtos.java
index 5b450b6..865e556 100644
--- a/protocol/src/main/java/org/apache/drill/exec/proto/SchemaExecProtos.java
+++ b/protocol/src/main/java/org/apache/drill/exec/proto/SchemaExecProtos.java
@@ -31,9 +31,9 @@ public final class SchemaExecProtos
         public static final org.apache.drill.exec.proto.SchemaExecProtos.FragmentHandle.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaExecProtos.FragmentHandle.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.ExecProtos.FragmentHandle>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.ExecProtos.FragmentHandle>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.ExecProtos.FragmentHandle message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.ExecProtos.FragmentHandle message) throws java.io.IOException
             {
                 if(message.hasQueryId())
                     output.writeObject(1, message.getQueryId(), org.apache.drill.exec.proto.SchemaUserBitShared.QueryId.WRITE, false);
@@ -71,12 +71,12 @@ public final class SchemaExecProtos
                 return org.apache.drill.exec.proto.ExecProtos.FragmentHandle.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.ExecProtos.FragmentHandle message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.ExecProtos.FragmentHandle message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.ExecProtos.FragmentHandle newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.ExecProtos.FragmentHandle.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.ExecProtos.FragmentHandle.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.ExecProtos.FragmentHandle.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.ExecProtos.FragmentHandle.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -132,7 +132,7 @@ public final class SchemaExecProtos
                 return org.apache.drill.exec.proto.ExecProtos.FragmentHandle.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.ExecProtos.FragmentHandle.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.ExecProtos.FragmentHandle.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -167,9 +167,9 @@ public final class SchemaExecProtos
         public static final org.apache.drill.exec.proto.SchemaExecProtos.ServerPreparedStatementState.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaExecProtos.ServerPreparedStatementState.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.ExecProtos.ServerPreparedStatementState>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.ExecProtos.ServerPreparedStatementState>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.ExecProtos.ServerPreparedStatementState message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.ExecProtos.ServerPreparedStatementState message) throws java.io.IOException
             {
                 if(message.hasSqlQuery())
                     output.writeString(1, message.getSqlQuery(), false);
@@ -199,12 +199,12 @@ public final class SchemaExecProtos
                 return org.apache.drill.exec.proto.ExecProtos.ServerPreparedStatementState.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.ExecProtos.ServerPreparedStatementState message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.ExecProtos.ServerPreparedStatementState message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.ExecProtos.ServerPreparedStatementState newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.ExecProtos.ServerPreparedStatementState.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.ExecProtos.ServerPreparedStatementState.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.ExecProtos.ServerPreparedStatementState.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.ExecProtos.ServerPreparedStatementState.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -249,7 +249,7 @@ public final class SchemaExecProtos
                 return org.apache.drill.exec.proto.ExecProtos.ServerPreparedStatementState.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.ExecProtos.ServerPreparedStatementState.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.ExecProtos.ServerPreparedStatementState.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/SchemaGeneralRPCProtos.java b/protocol/src/main/java/org/apache/drill/exec/proto/SchemaGeneralRPCProtos.java
index e6383c8..6892096 100644
--- a/protocol/src/main/java/org/apache/drill/exec/proto/SchemaGeneralRPCProtos.java
+++ b/protocol/src/main/java/org/apache/drill/exec/proto/SchemaGeneralRPCProtos.java
@@ -31,9 +31,9 @@ public final class SchemaGeneralRPCProtos
         public static final org.apache.drill.exec.proto.SchemaGeneralRPCProtos.Ack.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaGeneralRPCProtos.Ack.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.GeneralRPCProtos.Ack>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.GeneralRPCProtos.Ack>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.GeneralRPCProtos.Ack message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.GeneralRPCProtos.Ack message) throws java.io.IOException
             {
                 if(message.hasOk())
                     output.writeBool(1, message.getOk(), false);
@@ -63,12 +63,12 @@ public final class SchemaGeneralRPCProtos
                 return org.apache.drill.exec.proto.GeneralRPCProtos.Ack.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.GeneralRPCProtos.Ack message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.GeneralRPCProtos.Ack message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.GeneralRPCProtos.Ack newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.GeneralRPCProtos.Ack.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.GeneralRPCProtos.Ack.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.GeneralRPCProtos.Ack.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.GeneralRPCProtos.Ack.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -113,7 +113,7 @@ public final class SchemaGeneralRPCProtos
                 return org.apache.drill.exec.proto.GeneralRPCProtos.Ack.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.GeneralRPCProtos.Ack.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.GeneralRPCProtos.Ack.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -142,9 +142,9 @@ public final class SchemaGeneralRPCProtos
         public static final org.apache.drill.exec.proto.SchemaGeneralRPCProtos.RpcHeader.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaGeneralRPCProtos.RpcHeader.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.GeneralRPCProtos.RpcHeader>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.GeneralRPCProtos.RpcHeader>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.GeneralRPCProtos.RpcHeader message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.GeneralRPCProtos.RpcHeader message) throws java.io.IOException
             {
                 if(message.hasMode())
                     output.writeEnum(1, message.getMode().getNumber(), false);
@@ -178,12 +178,12 @@ public final class SchemaGeneralRPCProtos
                 return org.apache.drill.exec.proto.GeneralRPCProtos.RpcHeader.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.GeneralRPCProtos.RpcHeader message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.GeneralRPCProtos.RpcHeader message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.GeneralRPCProtos.RpcHeader newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.GeneralRPCProtos.RpcHeader.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.GeneralRPCProtos.RpcHeader.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.GeneralRPCProtos.RpcHeader.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.GeneralRPCProtos.RpcHeader.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -234,7 +234,7 @@ public final class SchemaGeneralRPCProtos
                 return org.apache.drill.exec.proto.GeneralRPCProtos.RpcHeader.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.GeneralRPCProtos.RpcHeader.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.GeneralRPCProtos.RpcHeader.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -267,9 +267,9 @@ public final class SchemaGeneralRPCProtos
         public static final org.apache.drill.exec.proto.SchemaGeneralRPCProtos.CompleteRpcMessage.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaGeneralRPCProtos.CompleteRpcMessage.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.GeneralRPCProtos.CompleteRpcMessage>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.GeneralRPCProtos.CompleteRpcMessage>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.GeneralRPCProtos.CompleteRpcMessage message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.GeneralRPCProtos.CompleteRpcMessage message) throws java.io.IOException
             {
                 if(message.hasHeader())
                     output.writeObject(1, message.getHeader(), org.apache.drill.exec.proto.SchemaGeneralRPCProtos.RpcHeader.WRITE, false);
@@ -306,12 +306,12 @@ public final class SchemaGeneralRPCProtos
                 return org.apache.drill.exec.proto.GeneralRPCProtos.CompleteRpcMessage.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.GeneralRPCProtos.CompleteRpcMessage message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.GeneralRPCProtos.CompleteRpcMessage message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.GeneralRPCProtos.CompleteRpcMessage newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.GeneralRPCProtos.CompleteRpcMessage.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.GeneralRPCProtos.CompleteRpcMessage.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.GeneralRPCProtos.CompleteRpcMessage.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.GeneralRPCProtos.CompleteRpcMessage.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -363,7 +363,7 @@ public final class SchemaGeneralRPCProtos
                 return org.apache.drill.exec.proto.GeneralRPCProtos.CompleteRpcMessage.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.GeneralRPCProtos.CompleteRpcMessage.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.GeneralRPCProtos.CompleteRpcMessage.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/SchemaUserBitShared.java b/protocol/src/main/java/org/apache/drill/exec/proto/SchemaUserBitShared.java
index 932872c..5aeff0f 100644
--- a/protocol/src/main/java/org/apache/drill/exec/proto/SchemaUserBitShared.java
+++ b/protocol/src/main/java/org/apache/drill/exec/proto/SchemaUserBitShared.java
@@ -31,9 +31,9 @@ public final class SchemaUserBitShared
         public static final org.apache.drill.exec.proto.SchemaUserBitShared.UserCredentials.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserBitShared.UserCredentials.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.UserCredentials>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.UserCredentials>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.UserCredentials message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.UserCredentials message) throws java.io.IOException
             {
                 if(message.hasUserName())
                     output.writeString(1, message.getUserName(), false);
@@ -63,12 +63,12 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.UserCredentials.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.UserCredentials message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.UserCredentials message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserBitShared.UserCredentials newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.UserCredentials.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.UserCredentials.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.UserCredentials.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.UserCredentials.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -113,7 +113,7 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.UserCredentials.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.UserCredentials.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.UserCredentials.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -142,9 +142,9 @@ public final class SchemaUserBitShared
         public static final org.apache.drill.exec.proto.SchemaUserBitShared.QueryId.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserBitShared.QueryId.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.QueryId>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.QueryId>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.QueryId message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.QueryId message) throws java.io.IOException
             {
                 if(message.hasPart1())
                     output.writeSFixed64(1, message.getPart1(), false);
@@ -176,12 +176,12 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.QueryId.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.QueryId message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.QueryId message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserBitShared.QueryId newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.QueryId.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.QueryId.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.QueryId.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.QueryId.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -229,7 +229,7 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.QueryId.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.QueryId.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.QueryId.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -260,9 +260,9 @@ public final class SchemaUserBitShared
         public static final org.apache.drill.exec.proto.SchemaUserBitShared.DrillPBError.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserBitShared.DrillPBError.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.DrillPBError>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.DrillPBError>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.DrillPBError message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.DrillPBError message) throws java.io.IOException
             {
                 if(message.hasErrorId())
                     output.writeString(1, message.getErrorId(), false);
@@ -305,12 +305,12 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.DrillPBError.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.DrillPBError message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.DrillPBError message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserBitShared.DrillPBError newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.DrillPBError.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.DrillPBError.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.DrillPBError.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.DrillPBError.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -373,7 +373,7 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.DrillPBError.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.DrillPBError.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.DrillPBError.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -412,9 +412,9 @@ public final class SchemaUserBitShared
         public static final org.apache.drill.exec.proto.SchemaUserBitShared.ExceptionWrapper.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserBitShared.ExceptionWrapper.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.ExceptionWrapper>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.ExceptionWrapper>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.ExceptionWrapper message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.ExceptionWrapper message) throws java.io.IOException
             {
                 if(message.hasExceptionClass())
                     output.writeString(1, message.getExceptionClass(), false);
@@ -452,12 +452,12 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.ExceptionWrapper.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.ExceptionWrapper message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.ExceptionWrapper message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserBitShared.ExceptionWrapper newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.ExceptionWrapper.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.ExceptionWrapper.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.ExceptionWrapper.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.ExceptionWrapper.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -513,7 +513,7 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.ExceptionWrapper.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.ExceptionWrapper.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.ExceptionWrapper.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -548,9 +548,9 @@ public final class SchemaUserBitShared
         public static final org.apache.drill.exec.proto.SchemaUserBitShared.StackTraceElementWrapper.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserBitShared.StackTraceElementWrapper.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.StackTraceElementWrapper>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.StackTraceElementWrapper>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.StackTraceElementWrapper message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.StackTraceElementWrapper message) throws java.io.IOException
             {
                 if(message.hasClassName())
                     output.writeString(1, message.getClassName(), false);
@@ -588,12 +588,12 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.StackTraceElementWrapper.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.StackTraceElementWrapper message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.StackTraceElementWrapper message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserBitShared.StackTraceElementWrapper newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.StackTraceElementWrapper.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.StackTraceElementWrapper.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.StackTraceElementWrapper.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.StackTraceElementWrapper.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -650,7 +650,7 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.StackTraceElementWrapper.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.StackTraceElementWrapper.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.StackTraceElementWrapper.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -687,9 +687,9 @@ public final class SchemaUserBitShared
         public static final org.apache.drill.exec.proto.SchemaUserBitShared.ParsingError.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserBitShared.ParsingError.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.ParsingError>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.ParsingError>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.ParsingError message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.ParsingError message) throws java.io.IOException
             {
                 if(message.hasStartColumn())
                     output.writeInt32(2, message.getStartColumn(), false);
@@ -725,12 +725,12 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.ParsingError.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.ParsingError message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.ParsingError message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserBitShared.ParsingError newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.ParsingError.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.ParsingError.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.ParsingError.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.ParsingError.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -784,7 +784,7 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.ParsingError.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.ParsingError.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.ParsingError.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -819,9 +819,9 @@ public final class SchemaUserBitShared
         public static final org.apache.drill.exec.proto.SchemaUserBitShared.RecordBatchDef.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserBitShared.RecordBatchDef.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.RecordBatchDef>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.RecordBatchDef>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.RecordBatchDef message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.RecordBatchDef message) throws java.io.IOException
             {
                 if(message.hasRecordCount())
                     output.writeInt32(1, message.getRecordCount(), false);
@@ -858,12 +858,12 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.RecordBatchDef.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.RecordBatchDef message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.RecordBatchDef message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserBitShared.RecordBatchDef newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.RecordBatchDef.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.RecordBatchDef.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.RecordBatchDef.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.RecordBatchDef.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -918,7 +918,7 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.RecordBatchDef.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.RecordBatchDef.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.RecordBatchDef.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -953,9 +953,9 @@ public final class SchemaUserBitShared
         public static final org.apache.drill.exec.proto.SchemaUserBitShared.NamePart.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserBitShared.NamePart.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.NamePart>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.NamePart>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.NamePart message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.NamePart message) throws java.io.IOException
             {
                 if(message.hasType())
                     output.writeEnum(1, message.getType().getNumber(), false);
@@ -990,12 +990,12 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.NamePart.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.NamePart message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.NamePart message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserBitShared.NamePart newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.NamePart.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.NamePart.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.NamePart.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.NamePart.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -1047,7 +1047,7 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.NamePart.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.NamePart.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.NamePart.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -1080,9 +1080,9 @@ public final class SchemaUserBitShared
         public static final org.apache.drill.exec.proto.SchemaUserBitShared.SerializedField.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserBitShared.SerializedField.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.SerializedField>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.SerializedField>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.SerializedField message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.SerializedField message) throws java.io.IOException
             {
                 if(message.hasMajorType())
                     output.writeObject(1, message.getMajorType(), org.apache.drill.common.types.SchemaTypeProtos.MajorType.WRITE, false);
@@ -1125,12 +1125,12 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.SerializedField.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.SerializedField message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.SerializedField message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserBitShared.SerializedField newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.SerializedField.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.SerializedField.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.SerializedField.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.SerializedField.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -1193,7 +1193,7 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.SerializedField.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.SerializedField.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.SerializedField.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -1232,9 +1232,9 @@ public final class SchemaUserBitShared
         public static final org.apache.drill.exec.proto.SchemaUserBitShared.NodeStatus.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserBitShared.NodeStatus.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.NodeStatus>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.NodeStatus>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.NodeStatus message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.NodeStatus message) throws java.io.IOException
             {
                 if(message.hasNodeId())
                     output.writeInt32(1, message.getNodeId(), false);
@@ -1266,12 +1266,12 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.NodeStatus.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.NodeStatus message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.NodeStatus message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserBitShared.NodeStatus newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.NodeStatus.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.NodeStatus.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.NodeStatus.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.NodeStatus.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -1319,7 +1319,7 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.NodeStatus.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.NodeStatus.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.NodeStatus.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -1350,9 +1350,9 @@ public final class SchemaUserBitShared
         public static final org.apache.drill.exec.proto.SchemaUserBitShared.QueryResult.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserBitShared.QueryResult.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.QueryResult>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.QueryResult>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.QueryResult message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.QueryResult message) throws java.io.IOException
             {
                 if(message.hasQueryState())
                     output.writeEnum(1, message.getQueryState().getNumber(), false);
@@ -1388,12 +1388,12 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.QueryResult.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.QueryResult message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.QueryResult message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserBitShared.QueryResult newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.QueryResult.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.QueryResult.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.QueryResult.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.QueryResult.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -1446,7 +1446,7 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.QueryResult.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.QueryResult.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.QueryResult.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -1479,9 +1479,9 @@ public final class SchemaUserBitShared
         public static final org.apache.drill.exec.proto.SchemaUserBitShared.QueryData.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserBitShared.QueryData.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.QueryData>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.QueryData>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.QueryData message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.QueryData message) throws java.io.IOException
             {
                 if(message.hasQueryId())
                     output.writeObject(1, message.getQueryId(), org.apache.drill.exec.proto.SchemaUserBitShared.QueryId.WRITE, false);
@@ -1519,12 +1519,12 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.QueryData.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.QueryData message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.QueryData message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserBitShared.QueryData newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.QueryData.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.QueryData.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.QueryData.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.QueryData.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -1580,7 +1580,7 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.QueryData.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.QueryData.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.QueryData.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -1615,9 +1615,9 @@ public final class SchemaUserBitShared
         public static final org.apache.drill.exec.proto.SchemaUserBitShared.QueryInfo.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserBitShared.QueryInfo.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.QueryInfo>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.QueryInfo>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.QueryInfo message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.QueryInfo message) throws java.io.IOException
             {
                 if(message.hasQuery())
                     output.writeString(1, message.getQuery(), false);
@@ -1662,12 +1662,12 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.QueryInfo.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.QueryInfo message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.QueryInfo message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserBitShared.QueryInfo newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.QueryInfo.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.QueryInfo.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.QueryInfo.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.QueryInfo.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -1734,7 +1734,7 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.QueryInfo.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.QueryInfo.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.QueryInfo.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -1777,9 +1777,9 @@ public final class SchemaUserBitShared
         public static final org.apache.drill.exec.proto.SchemaUserBitShared.QueryProfile.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserBitShared.QueryProfile.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.QueryProfile>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.QueryProfile>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.QueryProfile message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.QueryProfile message) throws java.io.IOException
             {
                 if(message.hasId())
                     output.writeObject(1, message.getId(), org.apache.drill.exec.proto.SchemaUserBitShared.QueryId.WRITE, false);
@@ -1856,12 +1856,12 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.QueryProfile.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.QueryProfile message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.QueryProfile message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserBitShared.QueryProfile newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.QueryProfile.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.QueryProfile.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.QueryProfile.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.QueryProfile.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -1975,7 +1975,7 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.QueryProfile.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.QueryProfile.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.QueryProfile.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -2048,9 +2048,9 @@ public final class SchemaUserBitShared
         public static final org.apache.drill.exec.proto.SchemaUserBitShared.MajorFragmentProfile.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserBitShared.MajorFragmentProfile.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.MajorFragmentProfile>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.MajorFragmentProfile>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.MajorFragmentProfile message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.MajorFragmentProfile message) throws java.io.IOException
             {
                 if(message.hasMajorFragmentId())
                     output.writeInt32(1, message.getMajorFragmentId(), false);
@@ -2083,12 +2083,12 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.MajorFragmentProfile.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.MajorFragmentProfile message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.MajorFragmentProfile message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserBitShared.MajorFragmentProfile newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.MajorFragmentProfile.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.MajorFragmentProfile.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.MajorFragmentProfile.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.MajorFragmentProfile.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -2137,7 +2137,7 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.MajorFragmentProfile.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.MajorFragmentProfile.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.MajorFragmentProfile.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -2168,9 +2168,9 @@ public final class SchemaUserBitShared
         public static final org.apache.drill.exec.proto.SchemaUserBitShared.MinorFragmentProfile.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserBitShared.MinorFragmentProfile.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.MinorFragmentProfile>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.MinorFragmentProfile>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.MinorFragmentProfile message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.MinorFragmentProfile message) throws java.io.IOException
             {
                 if(message.hasState())
                     output.writeEnum(1, message.getState().getNumber(), false);
@@ -2223,12 +2223,12 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.MinorFragmentProfile.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.MinorFragmentProfile message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.MinorFragmentProfile message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserBitShared.MinorFragmentProfile newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.MinorFragmentProfile.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.MinorFragmentProfile.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.MinorFragmentProfile.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.MinorFragmentProfile.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -2306,7 +2306,7 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.MinorFragmentProfile.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.MinorFragmentProfile.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.MinorFragmentProfile.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -2355,9 +2355,9 @@ public final class SchemaUserBitShared
         public static final org.apache.drill.exec.proto.SchemaUserBitShared.OperatorProfile.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserBitShared.OperatorProfile.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.OperatorProfile>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.OperatorProfile>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.OperatorProfile message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.OperatorProfile message) throws java.io.IOException
             {
                 for(org.apache.drill.exec.proto.UserBitShared.StreamProfile inputProfile : message.getInputProfileList())
                     output.writeObject(1, inputProfile, org.apache.drill.exec.proto.SchemaUserBitShared.StreamProfile.WRITE, true);
@@ -2403,12 +2403,12 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.OperatorProfile.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.OperatorProfile message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.OperatorProfile message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserBitShared.OperatorProfile newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.OperatorProfile.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.OperatorProfile.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.OperatorProfile.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.OperatorProfile.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -2476,7 +2476,7 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.OperatorProfile.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.OperatorProfile.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.OperatorProfile.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -2519,9 +2519,9 @@ public final class SchemaUserBitShared
         public static final org.apache.drill.exec.proto.SchemaUserBitShared.StreamProfile.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserBitShared.StreamProfile.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.StreamProfile>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.StreamProfile>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.StreamProfile message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.StreamProfile message) throws java.io.IOException
             {
                 if(message.hasRecords())
                     output.writeInt64(1, message.getRecords(), false);
@@ -2555,12 +2555,12 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.StreamProfile.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.StreamProfile message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.StreamProfile message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserBitShared.StreamProfile newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.StreamProfile.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.StreamProfile.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.StreamProfile.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.StreamProfile.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -2611,7 +2611,7 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.StreamProfile.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.StreamProfile.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.StreamProfile.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -2644,9 +2644,9 @@ public final class SchemaUserBitShared
         public static final org.apache.drill.exec.proto.SchemaUserBitShared.MetricValue.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserBitShared.MetricValue.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.MetricValue>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.MetricValue>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.MetricValue message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.MetricValue message) throws java.io.IOException
             {
                 if(message.hasMetricId())
                     output.writeInt32(1, message.getMetricId(), false);
@@ -2680,12 +2680,12 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.MetricValue.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.MetricValue message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.MetricValue message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserBitShared.MetricValue newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.MetricValue.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.MetricValue.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.MetricValue.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.MetricValue.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -2736,7 +2736,7 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.MetricValue.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.MetricValue.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.MetricValue.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -2769,9 +2769,9 @@ public final class SchemaUserBitShared
         public static final org.apache.drill.exec.proto.SchemaUserBitShared.Registry.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserBitShared.Registry.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.Registry>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.Registry>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.Registry message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.Registry message) throws java.io.IOException
             {
                 for(org.apache.drill.exec.proto.UserBitShared.Jar jar : message.getJarList())
                     output.writeObject(1, jar, org.apache.drill.exec.proto.SchemaUserBitShared.Jar.WRITE, true);
@@ -2802,12 +2802,12 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.Registry.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.Registry message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.Registry message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserBitShared.Registry newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.Registry.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.Registry.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.Registry.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.Registry.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -2853,7 +2853,7 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.Registry.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.Registry.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.Registry.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -2882,9 +2882,9 @@ public final class SchemaUserBitShared
         public static final org.apache.drill.exec.proto.SchemaUserBitShared.Jar.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserBitShared.Jar.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.Jar>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.Jar>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.Jar message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.Jar message) throws java.io.IOException
             {
                 if(message.hasName())
                     output.writeString(1, message.getName(), false);
@@ -2916,12 +2916,12 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.Jar.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.Jar message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.Jar message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserBitShared.Jar newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.Jar.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.Jar.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.Jar.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.Jar.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -2969,7 +2969,7 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.Jar.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.Jar.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.Jar.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -3000,9 +3000,9 @@ public final class SchemaUserBitShared
         public static final org.apache.drill.exec.proto.SchemaUserBitShared.SaslMessage.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserBitShared.SaslMessage.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.SaslMessage>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.SaslMessage>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.SaslMessage message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.SaslMessage message) throws java.io.IOException
             {
                 if(message.hasMechanism())
                     output.writeString(1, message.getMechanism(), false);
@@ -3037,12 +3037,12 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.SaslMessage.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.SaslMessage message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.SaslMessage message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserBitShared.SaslMessage newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.SaslMessage.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserBitShared.SaslMessage.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.SaslMessage.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserBitShared.SaslMessage.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -3093,7 +3093,7 @@ public final class SchemaUserBitShared
                 return org.apache.drill.exec.proto.UserBitShared.SaslMessage.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.SaslMessage.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserBitShared.SaslMessage.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/SchemaUserProtos.java b/protocol/src/main/java/org/apache/drill/exec/proto/SchemaUserProtos.java
index 7711c8f..543765d 100644
--- a/protocol/src/main/java/org/apache/drill/exec/proto/SchemaUserProtos.java
+++ b/protocol/src/main/java/org/apache/drill/exec/proto/SchemaUserProtos.java
@@ -31,9 +31,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.Property.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.Property.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.Property>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.Property>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.Property message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.Property message) throws java.io.IOException
             {
                 if(message.hasKey())
                     output.writeString(1, message.getKey(), false);
@@ -65,12 +65,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.Property.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.Property message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.Property message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.Property newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.Property.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.Property.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.Property.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.Property.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -118,7 +118,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.Property.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.Property.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.Property.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -149,9 +149,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.UserProperties.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.UserProperties.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.UserProperties>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.UserProperties>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.UserProperties message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.UserProperties message) throws java.io.IOException
             {
                 for(org.apache.drill.exec.proto.UserProtos.Property properties : message.getPropertiesList())
                     output.writeObject(1, properties, org.apache.drill.exec.proto.SchemaUserProtos.Property.WRITE, true);
@@ -182,12 +182,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.UserProperties.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.UserProperties message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.UserProperties message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.UserProperties newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.UserProperties.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.UserProperties.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.UserProperties.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.UserProperties.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -233,7 +233,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.UserProperties.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.UserProperties.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.UserProperties.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -262,9 +262,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.RpcEndpointInfos.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.RpcEndpointInfos.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.RpcEndpointInfos>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.RpcEndpointInfos>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.RpcEndpointInfos message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.RpcEndpointInfos message) throws java.io.IOException
             {
                 if(message.hasName())
                     output.writeString(1, message.getName(), false);
@@ -308,12 +308,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.RpcEndpointInfos.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.RpcEndpointInfos message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.RpcEndpointInfos message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.RpcEndpointInfos newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.RpcEndpointInfos.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.RpcEndpointInfos.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.RpcEndpointInfos.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.RpcEndpointInfos.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -379,7 +379,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.RpcEndpointInfos.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.RpcEndpointInfos.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.RpcEndpointInfos.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -422,9 +422,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.UserToBitHandshake.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.UserToBitHandshake.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.UserToBitHandshake>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.UserToBitHandshake>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.UserToBitHandshake message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.UserToBitHandshake message) throws java.io.IOException
             {
                 if(message.hasChannel())
                     output.writeEnum(1, message.getChannel().getNumber(), false);
@@ -473,12 +473,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.UserToBitHandshake.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.UserToBitHandshake message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.UserToBitHandshake message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.UserToBitHandshake newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.UserToBitHandshake.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.UserToBitHandshake.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.UserToBitHandshake.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.UserToBitHandshake.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -550,7 +550,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.UserToBitHandshake.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.UserToBitHandshake.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.UserToBitHandshake.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -595,9 +595,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.RequestResults.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.RequestResults.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.RequestResults>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.RequestResults>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.RequestResults message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.RequestResults message) throws java.io.IOException
             {
                 if(message.hasQueryId())
                     output.writeObject(1, message.getQueryId(), org.apache.drill.exec.proto.SchemaUserBitShared.QueryId.WRITE, false);
@@ -630,12 +630,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.RequestResults.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.RequestResults message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.RequestResults message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.RequestResults newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.RequestResults.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.RequestResults.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.RequestResults.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.RequestResults.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -684,7 +684,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.RequestResults.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.RequestResults.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.RequestResults.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -715,9 +715,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.GetQueryPlanFragments.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.GetQueryPlanFragments.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetQueryPlanFragments>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetQueryPlanFragments>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetQueryPlanFragments message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetQueryPlanFragments message) throws java.io.IOException
             {
                 if(message.hasQuery())
                     output.writeString(1, message.getQuery(), false);
@@ -751,12 +751,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.GetQueryPlanFragments.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetQueryPlanFragments message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetQueryPlanFragments message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.GetQueryPlanFragments newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetQueryPlanFragments.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetQueryPlanFragments.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetQueryPlanFragments.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetQueryPlanFragments.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -807,7 +807,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.GetQueryPlanFragments.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetQueryPlanFragments.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetQueryPlanFragments.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -840,9 +840,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.QueryPlanFragments.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.QueryPlanFragments.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.QueryPlanFragments>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.QueryPlanFragments>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.QueryPlanFragments message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.QueryPlanFragments message) throws java.io.IOException
             {
                 if(message.hasStatus())
                     output.writeEnum(1, message.getStatus().getNumber(), false);
@@ -881,12 +881,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.QueryPlanFragments.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.QueryPlanFragments message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.QueryPlanFragments message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.QueryPlanFragments newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.QueryPlanFragments.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.QueryPlanFragments.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.QueryPlanFragments.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.QueryPlanFragments.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -943,7 +943,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.QueryPlanFragments.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.QueryPlanFragments.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.QueryPlanFragments.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -978,9 +978,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.BitToUserHandshake.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.BitToUserHandshake.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.BitToUserHandshake>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.BitToUserHandshake>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.BitToUserHandshake message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.BitToUserHandshake message) throws java.io.IOException
             {
                 if(message.hasRpcVersion())
                     output.writeInt32(2, message.getRpcVersion(), false);
@@ -1027,12 +1027,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.BitToUserHandshake.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.BitToUserHandshake message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.BitToUserHandshake message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.BitToUserHandshake newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.BitToUserHandshake.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.BitToUserHandshake.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.BitToUserHandshake.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.BitToUserHandshake.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -1102,7 +1102,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.BitToUserHandshake.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.BitToUserHandshake.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.BitToUserHandshake.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -1147,9 +1147,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.LikeFilter.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.LikeFilter.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.LikeFilter>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.LikeFilter>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.LikeFilter message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.LikeFilter message) throws java.io.IOException
             {
                 if(message.hasPattern())
                     output.writeString(1, message.getPattern(), false);
@@ -1181,12 +1181,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.LikeFilter.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.LikeFilter message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.LikeFilter message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.LikeFilter newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.LikeFilter.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.LikeFilter.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.LikeFilter.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.LikeFilter.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -1234,7 +1234,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.LikeFilter.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.LikeFilter.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.LikeFilter.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -1265,9 +1265,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.GetCatalogsReq.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.GetCatalogsReq.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetCatalogsReq>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetCatalogsReq>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetCatalogsReq message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetCatalogsReq message) throws java.io.IOException
             {
                 if(message.hasCatalogNameFilter())
                     output.writeObject(1, message.getCatalogNameFilter(), org.apache.drill.exec.proto.SchemaUserProtos.LikeFilter.WRITE, false);
@@ -1298,12 +1298,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.GetCatalogsReq.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetCatalogsReq message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetCatalogsReq message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.GetCatalogsReq newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetCatalogsReq.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetCatalogsReq.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetCatalogsReq.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetCatalogsReq.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -1349,7 +1349,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.GetCatalogsReq.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetCatalogsReq.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetCatalogsReq.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -1378,9 +1378,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.CatalogMetadata.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.CatalogMetadata.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.CatalogMetadata>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.CatalogMetadata>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.CatalogMetadata message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.CatalogMetadata message) throws java.io.IOException
             {
                 if(message.hasCatalogName())
                     output.writeString(1, message.getCatalogName(), false);
@@ -1414,12 +1414,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.CatalogMetadata.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.CatalogMetadata message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.CatalogMetadata message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.CatalogMetadata newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.CatalogMetadata.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.CatalogMetadata.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.CatalogMetadata.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.CatalogMetadata.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -1470,7 +1470,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.CatalogMetadata.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.CatalogMetadata.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.CatalogMetadata.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -1503,9 +1503,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.GetCatalogsResp.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.GetCatalogsResp.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetCatalogsResp>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetCatalogsResp>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetCatalogsResp message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetCatalogsResp message) throws java.io.IOException
             {
                 if(message.hasStatus())
                     output.writeEnum(1, message.getStatus().getNumber(), false);
@@ -1541,12 +1541,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.GetCatalogsResp.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetCatalogsResp message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetCatalogsResp message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.GetCatalogsResp newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetCatalogsResp.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetCatalogsResp.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetCatalogsResp.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetCatalogsResp.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -1599,7 +1599,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.GetCatalogsResp.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetCatalogsResp.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetCatalogsResp.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -1632,9 +1632,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.GetSchemasReq.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.GetSchemasReq.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetSchemasReq>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetSchemasReq>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetSchemasReq message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetSchemasReq message) throws java.io.IOException
             {
                 if(message.hasCatalogNameFilter())
                     output.writeObject(1, message.getCatalogNameFilter(), org.apache.drill.exec.proto.SchemaUserProtos.LikeFilter.WRITE, false);
@@ -1668,12 +1668,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.GetSchemasReq.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetSchemasReq message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetSchemasReq message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.GetSchemasReq newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetSchemasReq.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetSchemasReq.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetSchemasReq.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetSchemasReq.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -1723,7 +1723,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.GetSchemasReq.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetSchemasReq.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetSchemasReq.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -1754,9 +1754,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.SchemaMetadata.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.SchemaMetadata.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.SchemaMetadata>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.SchemaMetadata>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.SchemaMetadata message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.SchemaMetadata message) throws java.io.IOException
             {
                 if(message.hasCatalogName())
                     output.writeString(1, message.getCatalogName(), false);
@@ -1794,12 +1794,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.SchemaMetadata.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.SchemaMetadata message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.SchemaMetadata message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.SchemaMetadata newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.SchemaMetadata.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.SchemaMetadata.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.SchemaMetadata.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.SchemaMetadata.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -1856,7 +1856,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.SchemaMetadata.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.SchemaMetadata.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.SchemaMetadata.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -1893,9 +1893,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.GetSchemasResp.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.GetSchemasResp.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetSchemasResp>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetSchemasResp>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetSchemasResp message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetSchemasResp message) throws java.io.IOException
             {
                 if(message.hasStatus())
                     output.writeEnum(1, message.getStatus().getNumber(), false);
@@ -1931,12 +1931,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.GetSchemasResp.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetSchemasResp message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetSchemasResp message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.GetSchemasResp newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetSchemasResp.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetSchemasResp.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetSchemasResp.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetSchemasResp.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -1989,7 +1989,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.GetSchemasResp.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetSchemasResp.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetSchemasResp.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -2022,9 +2022,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.GetTablesReq.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.GetTablesReq.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetTablesReq>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetTablesReq>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetTablesReq message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetTablesReq message) throws java.io.IOException
             {
                 if(message.hasCatalogNameFilter())
                     output.writeObject(1, message.getCatalogNameFilter(), org.apache.drill.exec.proto.SchemaUserProtos.LikeFilter.WRITE, false);
@@ -2063,12 +2063,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.GetTablesReq.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetTablesReq message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetTablesReq message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.GetTablesReq newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetTablesReq.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetTablesReq.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetTablesReq.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetTablesReq.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -2125,7 +2125,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.GetTablesReq.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetTablesReq.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetTablesReq.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -2160,9 +2160,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.TableMetadata.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.TableMetadata.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.TableMetadata>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.TableMetadata>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.TableMetadata message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.TableMetadata message) throws java.io.IOException
             {
                 if(message.hasCatalogName())
                     output.writeString(1, message.getCatalogName(), false);
@@ -2198,12 +2198,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.TableMetadata.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.TableMetadata message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.TableMetadata message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.TableMetadata newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.TableMetadata.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.TableMetadata.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.TableMetadata.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.TableMetadata.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -2257,7 +2257,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.TableMetadata.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.TableMetadata.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.TableMetadata.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -2292,9 +2292,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.GetTablesResp.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.GetTablesResp.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetTablesResp>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetTablesResp>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetTablesResp message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetTablesResp message) throws java.io.IOException
             {
                 if(message.hasStatus())
                     output.writeEnum(1, message.getStatus().getNumber(), false);
@@ -2330,12 +2330,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.GetTablesResp.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetTablesResp message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetTablesResp message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.GetTablesResp newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetTablesResp.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetTablesResp.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetTablesResp.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetTablesResp.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -2388,7 +2388,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.GetTablesResp.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetTablesResp.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetTablesResp.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -2421,9 +2421,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.GetColumnsReq.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.GetColumnsReq.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetColumnsReq>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetColumnsReq>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetColumnsReq message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetColumnsReq message) throws java.io.IOException
             {
                 if(message.hasCatalogNameFilter())
                     output.writeObject(1, message.getCatalogNameFilter(), org.apache.drill.exec.proto.SchemaUserProtos.LikeFilter.WRITE, false);
@@ -2463,12 +2463,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.GetColumnsReq.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetColumnsReq message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetColumnsReq message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.GetColumnsReq newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetColumnsReq.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetColumnsReq.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetColumnsReq.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetColumnsReq.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -2526,7 +2526,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.GetColumnsReq.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetColumnsReq.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetColumnsReq.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -2561,9 +2561,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.ColumnMetadata.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.ColumnMetadata.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.ColumnMetadata>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.ColumnMetadata>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.ColumnMetadata message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.ColumnMetadata message) throws java.io.IOException
             {
                 if(message.hasCatalogName())
                     output.writeString(1, message.getCatalogName(), false);
@@ -2625,12 +2625,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.ColumnMetadata.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.ColumnMetadata message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.ColumnMetadata message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.ColumnMetadata newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.ColumnMetadata.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.ColumnMetadata.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.ColumnMetadata.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.ColumnMetadata.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -2723,7 +2723,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.ColumnMetadata.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.ColumnMetadata.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.ColumnMetadata.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -2784,9 +2784,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.GetColumnsResp.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.GetColumnsResp.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetColumnsResp>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetColumnsResp>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetColumnsResp message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetColumnsResp message) throws java.io.IOException
             {
                 if(message.hasStatus())
                     output.writeEnum(1, message.getStatus().getNumber(), false);
@@ -2822,12 +2822,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.GetColumnsResp.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetColumnsResp message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetColumnsResp message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.GetColumnsResp newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetColumnsResp.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetColumnsResp.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetColumnsResp.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetColumnsResp.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -2880,7 +2880,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.GetColumnsResp.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetColumnsResp.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetColumnsResp.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -2913,9 +2913,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.CreatePreparedStatementReq.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.CreatePreparedStatementReq.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementReq>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementReq>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementReq message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementReq message) throws java.io.IOException
             {
                 if(message.hasSqlQuery())
                     output.writeString(1, message.getSqlQuery(), false);
@@ -2945,12 +2945,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementReq.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementReq message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementReq message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementReq newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementReq.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementReq.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementReq.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementReq.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -2995,7 +2995,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementReq.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementReq.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementReq.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -3024,9 +3024,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.ResultColumnMetadata.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.ResultColumnMetadata.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.ResultColumnMetadata>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.ResultColumnMetadata>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.ResultColumnMetadata message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.ResultColumnMetadata message) throws java.io.IOException
             {
                 if(message.hasCatalogName())
                     output.writeString(1, message.getCatalogName(), false);
@@ -3092,12 +3092,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.ResultColumnMetadata.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.ResultColumnMetadata message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.ResultColumnMetadata message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.ResultColumnMetadata newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.ResultColumnMetadata.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.ResultColumnMetadata.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.ResultColumnMetadata.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.ResultColumnMetadata.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -3196,7 +3196,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.ResultColumnMetadata.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.ResultColumnMetadata.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.ResultColumnMetadata.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -3261,9 +3261,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.PreparedStatementHandle.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.PreparedStatementHandle.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.PreparedStatementHandle>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.PreparedStatementHandle>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.PreparedStatementHandle message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.PreparedStatementHandle message) throws java.io.IOException
             {
                 if(message.hasServerInfo())
                     output.writeByteArray(1, message.getServerInfo().toByteArray(), false);
@@ -3294,12 +3294,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.PreparedStatementHandle.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.PreparedStatementHandle message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.PreparedStatementHandle message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.PreparedStatementHandle newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.PreparedStatementHandle.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.PreparedStatementHandle.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.PreparedStatementHandle.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.PreparedStatementHandle.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -3344,7 +3344,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.PreparedStatementHandle.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.PreparedStatementHandle.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.PreparedStatementHandle.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -3373,9 +3373,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.PreparedStatement.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.PreparedStatement.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.PreparedStatement>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.PreparedStatement>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.PreparedStatement message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.PreparedStatement message) throws java.io.IOException
             {
                 for(org.apache.drill.exec.proto.UserProtos.ResultColumnMetadata columns : message.getColumnsList())
                     output.writeObject(1, columns, org.apache.drill.exec.proto.SchemaUserProtos.ResultColumnMetadata.WRITE, true);
@@ -3409,12 +3409,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.PreparedStatement.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.PreparedStatement message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.PreparedStatement message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.PreparedStatement newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.PreparedStatement.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.PreparedStatement.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.PreparedStatement.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.PreparedStatement.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -3464,7 +3464,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.PreparedStatement.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.PreparedStatement.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.PreparedStatement.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -3495,9 +3495,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.CreatePreparedStatementResp.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.CreatePreparedStatementResp.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementResp>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementResp>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementResp message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementResp message) throws java.io.IOException
             {
                 if(message.hasStatus())
                     output.writeEnum(1, message.getStatus().getNumber(), false);
@@ -3533,12 +3533,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementResp.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementResp message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementResp message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementResp newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementResp.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementResp.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementResp.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementResp.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -3591,7 +3591,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementResp.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementResp.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.CreatePreparedStatementResp.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -3624,9 +3624,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.GetServerMetaReq.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.GetServerMetaReq.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetServerMetaReq>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetServerMetaReq>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetServerMetaReq message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetServerMetaReq message) throws java.io.IOException
             {
             }
             public boolean isInitialized(org.apache.drill.exec.proto.UserProtos.GetServerMetaReq message)
@@ -3654,12 +3654,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.GetServerMetaReq.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetServerMetaReq message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetServerMetaReq message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.GetServerMetaReq newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetServerMetaReq.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetServerMetaReq.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetServerMetaReq.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetServerMetaReq.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -3701,7 +3701,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.GetServerMetaReq.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetServerMetaReq.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetServerMetaReq.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -3728,9 +3728,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.ConvertSupport.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.ConvertSupport.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.ConvertSupport>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.ConvertSupport>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.ConvertSupport message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.ConvertSupport message) throws java.io.IOException
             {
                 if(message.hasFrom())
                     output.writeEnum(1, message.getFrom().getNumber(), false);
@@ -3762,12 +3762,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.ConvertSupport.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.ConvertSupport message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.ConvertSupport message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.ConvertSupport newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.ConvertSupport.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.ConvertSupport.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.ConvertSupport.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.ConvertSupport.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -3815,7 +3815,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.ConvertSupport.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.ConvertSupport.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.ConvertSupport.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -3846,9 +3846,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.GetServerMetaResp.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.GetServerMetaResp.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetServerMetaResp>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetServerMetaResp>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetServerMetaResp message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetServerMetaResp message) throws java.io.IOException
             {
                 if(message.hasStatus())
                     output.writeEnum(1, message.getStatus().getNumber(), false);
@@ -3884,12 +3884,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.GetServerMetaResp.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetServerMetaResp message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetServerMetaResp message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.GetServerMetaResp newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetServerMetaResp.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.GetServerMetaResp.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetServerMetaResp.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.GetServerMetaResp.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -3942,7 +3942,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.GetServerMetaResp.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetServerMetaResp.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.GetServerMetaResp.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -3975,9 +3975,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.ServerMeta.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.ServerMeta.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.ServerMeta>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.ServerMeta>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.ServerMeta message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.ServerMeta message) throws java.io.IOException
             {
                 if(message.hasAllTablesSelectable())
                     output.writeBool(1, message.getAllTablesSelectable(), false);
@@ -4106,12 +4106,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.ServerMeta.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.ServerMeta message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.ServerMeta message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.ServerMeta newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.ServerMeta.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.ServerMeta.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.ServerMeta.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.ServerMeta.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -4304,7 +4304,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.ServerMeta.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.ServerMeta.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.ServerMeta.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
@@ -4431,9 +4431,9 @@ public final class SchemaUserProtos
         public static final org.apache.drill.exec.proto.SchemaUserProtos.RunQuery.BuilderSchema MERGE =
             new org.apache.drill.exec.proto.SchemaUserProtos.RunQuery.BuilderSchema();
         
-        public static class MessageSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.RunQuery>
+        public static class MessageSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.RunQuery>
         {
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.RunQuery message) throws java.io.IOException
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.RunQuery message) throws java.io.IOException
             {
                 if(message.hasResultsMode())
                     output.writeEnum(1, message.getResultsMode().getNumber(), false);
@@ -4475,12 +4475,12 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.RunQuery.class.getName();
             }
             //unused
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.RunQuery message) throws java.io.IOException {}
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.RunQuery message) throws java.io.IOException {}
             public org.apache.drill.exec.proto.UserProtos.RunQuery newMessage() { return null; }
         }
-        public static class BuilderSchema implements com.dyuproject.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.RunQuery.Builder>
+        public static class BuilderSchema implements io.protostuff.Schema<org.apache.drill.exec.proto.UserProtos.RunQuery.Builder>
         {
-            public void mergeFrom(com.dyuproject.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.RunQuery.Builder builder) throws java.io.IOException
+            public void mergeFrom(io.protostuff.Input input, org.apache.drill.exec.proto.UserProtos.RunQuery.Builder builder) throws java.io.IOException
             {
                 for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
                 {
@@ -4542,7 +4542,7 @@ public final class SchemaUserProtos
                 return org.apache.drill.exec.proto.UserProtos.RunQuery.class.getName();
             }
             //unused
-            public void writeTo(com.dyuproject.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.RunQuery.Builder builder) throws java.io.IOException {}
+            public void writeTo(io.protostuff.Output output, org.apache.drill.exec.proto.UserProtos.RunQuery.Builder builder) throws java.io.IOException {}
         }
         public static java.lang.String getFieldName(int number)
         {
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/Ack.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/Ack.java
deleted file mode 100644
index 245aba1..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/Ack.java
+++ /dev/null
@@ -1,163 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class Ack implements Externalizable, Message<Ack>, Schema<Ack>
-{
-
-    public static Schema<Ack> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static Ack getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final Ack DEFAULT_INSTANCE = new Ack();
-
-    
-    private Boolean ok;
-
-    public Ack()
-    {
-        
-    }
-
-    // getters and setters
-
-    // ok
-
-    public Boolean getOk()
-    {
-        return ok;
-    }
-
-    public Ack setOk(Boolean ok)
-    {
-        this.ok = ok;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<Ack> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public Ack newMessage()
-    {
-        return new Ack();
-    }
-
-    public Class<Ack> typeClass()
-    {
-        return Ack.class;
-    }
-
-    public String messageName()
-    {
-        return Ack.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return Ack.class.getName();
-    }
-
-    public boolean isInitialized(Ack message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, Ack message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.ok = input.readBool();
-                    break;
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, Ack message) throws IOException
-    {
-        if(message.ok != null)
-            output.writeBool(1, message.ok, false);
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "ok";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("ok", 1);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/BitClientHandshake.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/BitClientHandshake.java
deleted file mode 100644
index 3a27f70..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/BitClientHandshake.java
+++ /dev/null
@@ -1,185 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class BitClientHandshake implements Externalizable, Message<BitClientHandshake>, Schema<BitClientHandshake>
-{
-
-    public static Schema<BitClientHandshake> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static BitClientHandshake getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final BitClientHandshake DEFAULT_INSTANCE = new BitClientHandshake();
-
-    
-    private int rpcVersion;
-    private RpcChannel channel;
-
-    public BitClientHandshake()
-    {
-        
-    }
-
-    // getters and setters
-
-    // rpcVersion
-
-    public int getRpcVersion()
-    {
-        return rpcVersion;
-    }
-
-    public BitClientHandshake setRpcVersion(int rpcVersion)
-    {
-        this.rpcVersion = rpcVersion;
-        return this;
-    }
-
-    // channel
-
-    public RpcChannel getChannel()
-    {
-        return channel == null ? RpcChannel.BIT_DATA : channel;
-    }
-
-    public BitClientHandshake setChannel(RpcChannel channel)
-    {
-        this.channel = channel;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<BitClientHandshake> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public BitClientHandshake newMessage()
-    {
-        return new BitClientHandshake();
-    }
-
-    public Class<BitClientHandshake> typeClass()
-    {
-        return BitClientHandshake.class;
-    }
-
-    public String messageName()
-    {
-        return BitClientHandshake.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return BitClientHandshake.class.getName();
-    }
-
-    public boolean isInitialized(BitClientHandshake message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, BitClientHandshake message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.rpcVersion = input.readInt32();
-                    break;
-                case 2:
-                    message.channel = RpcChannel.valueOf(input.readEnum());
-                    break;
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, BitClientHandshake message) throws IOException
-    {
-        if(message.rpcVersion != 0)
-            output.writeInt32(1, message.rpcVersion, false);
-
-        if(message.channel != null)
-             output.writeEnum(2, message.channel.number, false);
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "rpcVersion";
-            case 2: return "channel";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("rpcVersion", 1);
-        __fieldMap.put("channel", 2);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/BitControlHandshake.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/BitControlHandshake.java
deleted file mode 100644
index fa6ef82..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/BitControlHandshake.java
+++ /dev/null
@@ -1,241 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-import java.util.ArrayList;
-import java.util.List;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class BitControlHandshake implements Externalizable, Message<BitControlHandshake>, Schema<BitControlHandshake>
-{
-
-    public static Schema<BitControlHandshake> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static BitControlHandshake getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final BitControlHandshake DEFAULT_INSTANCE = new BitControlHandshake();
-
-    
-    private int rpcVersion;
-    private RpcChannel channel;
-    private DrillbitEndpoint endpoint;
-    private List<String> authenticationMechanisms;
-
-    public BitControlHandshake()
-    {
-        
-    }
-
-    // getters and setters
-
-    // rpcVersion
-
-    public int getRpcVersion()
-    {
-        return rpcVersion;
-    }
-
-    public BitControlHandshake setRpcVersion(int rpcVersion)
-    {
-        this.rpcVersion = rpcVersion;
-        return this;
-    }
-
-    // channel
-
-    public RpcChannel getChannel()
-    {
-        return channel == null ? RpcChannel.BIT_CONTROL : channel;
-    }
-
-    public BitControlHandshake setChannel(RpcChannel channel)
-    {
-        this.channel = channel;
-        return this;
-    }
-
-    // endpoint
-
-    public DrillbitEndpoint getEndpoint()
-    {
-        return endpoint;
-    }
-
-    public BitControlHandshake setEndpoint(DrillbitEndpoint endpoint)
-    {
-        this.endpoint = endpoint;
-        return this;
-    }
-
-    // authenticationMechanisms
-
-    public List<String> getAuthenticationMechanismsList()
-    {
-        return authenticationMechanisms;
-    }
-
-    public BitControlHandshake setAuthenticationMechanismsList(List<String> authenticationMechanisms)
-    {
-        this.authenticationMechanisms = authenticationMechanisms;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<BitControlHandshake> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public BitControlHandshake newMessage()
-    {
-        return new BitControlHandshake();
-    }
-
-    public Class<BitControlHandshake> typeClass()
-    {
-        return BitControlHandshake.class;
-    }
-
-    public String messageName()
-    {
-        return BitControlHandshake.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return BitControlHandshake.class.getName();
-    }
-
-    public boolean isInitialized(BitControlHandshake message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, BitControlHandshake message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.rpcVersion = input.readInt32();
-                    break;
-                case 2:
-                    message.channel = RpcChannel.valueOf(input.readEnum());
-                    break;
-                case 3:
-                    message.endpoint = input.mergeObject(message.endpoint, DrillbitEndpoint.getSchema());
-                    break;
-
-                case 4:
-                    if(message.authenticationMechanisms == null)
-                        message.authenticationMechanisms = new ArrayList<String>();
-                    message.authenticationMechanisms.add(input.readString());
-                    break;
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, BitControlHandshake message) throws IOException
-    {
-        if(message.rpcVersion != 0)
-            output.writeInt32(1, message.rpcVersion, false);
-
-        if(message.channel != null)
-             output.writeEnum(2, message.channel.number, false);
-
-        if(message.endpoint != null)
-             output.writeObject(3, message.endpoint, DrillbitEndpoint.getSchema(), false);
-
-
-        if(message.authenticationMechanisms != null)
-        {
-            for(String authenticationMechanisms : message.authenticationMechanisms)
-            {
-                if(authenticationMechanisms != null)
-                    output.writeString(4, authenticationMechanisms, true);
-            }
-        }
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "rpcVersion";
-            case 2: return "channel";
-            case 3: return "endpoint";
-            case 4: return "authenticationMechanisms";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("rpcVersion", 1);
-        __fieldMap.put("channel", 2);
-        __fieldMap.put("endpoint", 3);
-        __fieldMap.put("authenticationMechanisms", 4);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/BitServerHandshake.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/BitServerHandshake.java
deleted file mode 100644
index b6a18c4..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/BitServerHandshake.java
+++ /dev/null
@@ -1,195 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-import java.util.ArrayList;
-import java.util.List;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class BitServerHandshake implements Externalizable, Message<BitServerHandshake>, Schema<BitServerHandshake>
-{
-
-    public static Schema<BitServerHandshake> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static BitServerHandshake getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final BitServerHandshake DEFAULT_INSTANCE = new BitServerHandshake();
-
-    
-    private int rpcVersion;
-    private List<String> authenticationMechanisms;
-
-    public BitServerHandshake()
-    {
-        
-    }
-
-    // getters and setters
-
-    // rpcVersion
-
-    public int getRpcVersion()
-    {
-        return rpcVersion;
-    }
-
-    public BitServerHandshake setRpcVersion(int rpcVersion)
-    {
-        this.rpcVersion = rpcVersion;
-        return this;
-    }
-
-    // authenticationMechanisms
-
-    public List<String> getAuthenticationMechanismsList()
-    {
-        return authenticationMechanisms;
-    }
-
-    public BitServerHandshake setAuthenticationMechanismsList(List<String> authenticationMechanisms)
-    {
-        this.authenticationMechanisms = authenticationMechanisms;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<BitServerHandshake> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public BitServerHandshake newMessage()
-    {
-        return new BitServerHandshake();
-    }
-
-    public Class<BitServerHandshake> typeClass()
-    {
-        return BitServerHandshake.class;
-    }
-
-    public String messageName()
-    {
-        return BitServerHandshake.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return BitServerHandshake.class.getName();
-    }
-
-    public boolean isInitialized(BitServerHandshake message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, BitServerHandshake message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.rpcVersion = input.readInt32();
-                    break;
-                case 2:
-                    if(message.authenticationMechanisms == null)
-                        message.authenticationMechanisms = new ArrayList<String>();
-                    message.authenticationMechanisms.add(input.readString());
-                    break;
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, BitServerHandshake message) throws IOException
-    {
-        if(message.rpcVersion != 0)
-            output.writeInt32(1, message.rpcVersion, false);
-
-        if(message.authenticationMechanisms != null)
-        {
-            for(String authenticationMechanisms : message.authenticationMechanisms)
-            {
-                if(authenticationMechanisms != null)
-                    output.writeString(2, authenticationMechanisms, true);
-            }
-        }
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "rpcVersion";
-            case 2: return "authenticationMechanisms";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("rpcVersion", 1);
-        __fieldMap.put("authenticationMechanisms", 2);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/BitStatus.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/BitStatus.java
deleted file mode 100644
index 0cba0c8..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/BitStatus.java
+++ /dev/null
@@ -1,175 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-import java.util.ArrayList;
-import java.util.List;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class BitStatus implements Externalizable, Message<BitStatus>, Schema<BitStatus>
-{
-
-    public static Schema<BitStatus> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static BitStatus getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final BitStatus DEFAULT_INSTANCE = new BitStatus();
-
-    
-    private List<FragmentStatus> fragmentStatus;
-
-    public BitStatus()
-    {
-        
-    }
-
-    // getters and setters
-
-    // fragmentStatus
-
-    public List<FragmentStatus> getFragmentStatusList()
-    {
-        return fragmentStatus;
-    }
-
-    public BitStatus setFragmentStatusList(List<FragmentStatus> fragmentStatus)
-    {
-        this.fragmentStatus = fragmentStatus;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<BitStatus> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public BitStatus newMessage()
-    {
-        return new BitStatus();
-    }
-
-    public Class<BitStatus> typeClass()
-    {
-        return BitStatus.class;
-    }
-
-    public String messageName()
-    {
-        return BitStatus.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return BitStatus.class.getName();
-    }
-
-    public boolean isInitialized(BitStatus message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, BitStatus message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    if(message.fragmentStatus == null)
-                        message.fragmentStatus = new ArrayList<FragmentStatus>();
-                    message.fragmentStatus.add(input.mergeObject(null, FragmentStatus.getSchema()));
-                    break;
-
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, BitStatus message) throws IOException
-    {
-        if(message.fragmentStatus != null)
-        {
-            for(FragmentStatus fragmentStatus : message.fragmentStatus)
-            {
-                if(fragmentStatus != null)
-                    output.writeObject(1, fragmentStatus, FragmentStatus.getSchema(), true);
-            }
-        }
-
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "fragmentStatus";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("fragmentStatus", 1);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/BitToUserHandshake.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/BitToUserHandshake.java
deleted file mode 100644
index 9c87c61..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/BitToUserHandshake.java
+++ /dev/null
@@ -1,359 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-import java.util.ArrayList;
-import java.util.List;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class BitToUserHandshake implements Externalizable, Message<BitToUserHandshake>, Schema<BitToUserHandshake>
-{
-
-    public static Schema<BitToUserHandshake> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static BitToUserHandshake getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final BitToUserHandshake DEFAULT_INSTANCE = new BitToUserHandshake();
-
-    
-    private int rpcVersion;
-    private HandshakeStatus status;
-    private String errorId;
-    private String errorMessage;
-    private RpcEndpointInfos serverInfos;
-    private List<String> authenticationMechanisms;
-    private List<RpcType> supportedMethods;
-    private Boolean encrypted;
-    private int maxWrappedSize;
-
-    public BitToUserHandshake()
-    {
-        
-    }
-
-    // getters and setters
-
-    // rpcVersion
-
-    public int getRpcVersion()
-    {
-        return rpcVersion;
-    }
-
-    public BitToUserHandshake setRpcVersion(int rpcVersion)
-    {
-        this.rpcVersion = rpcVersion;
-        return this;
-    }
-
-    // status
-
-    public HandshakeStatus getStatus()
-    {
-        return status == null ? HandshakeStatus.SUCCESS : status;
-    }
-
-    public BitToUserHandshake setStatus(HandshakeStatus status)
-    {
-        this.status = status;
-        return this;
-    }
-
-    // errorId
-
-    public String getErrorId()
-    {
-        return errorId;
-    }
-
-    public BitToUserHandshake setErrorId(String errorId)
-    {
-        this.errorId = errorId;
-        return this;
-    }
-
-    // errorMessage
-
-    public String getErrorMessage()
-    {
-        return errorMessage;
-    }
-
-    public BitToUserHandshake setErrorMessage(String errorMessage)
-    {
-        this.errorMessage = errorMessage;
-        return this;
-    }
-
-    // serverInfos
-
-    public RpcEndpointInfos getServerInfos()
-    {
-        return serverInfos;
-    }
-
-    public BitToUserHandshake setServerInfos(RpcEndpointInfos serverInfos)
-    {
-        this.serverInfos = serverInfos;
-        return this;
-    }
-
-    // authenticationMechanisms
-
-    public List<String> getAuthenticationMechanismsList()
-    {
-        return authenticationMechanisms;
-    }
-
-    public BitToUserHandshake setAuthenticationMechanismsList(List<String> authenticationMechanisms)
-    {
-        this.authenticationMechanisms = authenticationMechanisms;
-        return this;
-    }
-
-    // supportedMethods
-
-    public List<RpcType> getSupportedMethodsList()
-    {
-        return supportedMethods;
-    }
-
-    public BitToUserHandshake setSupportedMethodsList(List<RpcType> supportedMethods)
-    {
-        this.supportedMethods = supportedMethods;
-        return this;
-    }
-
-    // encrypted
-
-    public Boolean getEncrypted()
-    {
-        return encrypted;
-    }
-
-    public BitToUserHandshake setEncrypted(Boolean encrypted)
-    {
-        this.encrypted = encrypted;
-        return this;
-    }
-
-    // maxWrappedSize
-
-    public int getMaxWrappedSize()
-    {
-        return maxWrappedSize;
-    }
-
-    public BitToUserHandshake setMaxWrappedSize(int maxWrappedSize)
-    {
-        this.maxWrappedSize = maxWrappedSize;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<BitToUserHandshake> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public BitToUserHandshake newMessage()
-    {
-        return new BitToUserHandshake();
-    }
-
-    public Class<BitToUserHandshake> typeClass()
-    {
-        return BitToUserHandshake.class;
-    }
-
-    public String messageName()
-    {
-        return BitToUserHandshake.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return BitToUserHandshake.class.getName();
-    }
-
-    public boolean isInitialized(BitToUserHandshake message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, BitToUserHandshake message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 2:
-                    message.rpcVersion = input.readInt32();
-                    break;
-                case 3:
-                    message.status = HandshakeStatus.valueOf(input.readEnum());
-                    break;
-                case 4:
-                    message.errorId = input.readString();
-                    break;
-                case 5:
-                    message.errorMessage = input.readString();
-                    break;
-                case 6:
-                    message.serverInfos = input.mergeObject(message.serverInfos, RpcEndpointInfos.getSchema());
-                    break;
-
-                case 7:
-                    if(message.authenticationMechanisms == null)
-                        message.authenticationMechanisms = new ArrayList<String>();
-                    message.authenticationMechanisms.add(input.readString());
-                    break;
-                case 8:
-                    if(message.supportedMethods == null)
-                        message.supportedMethods = new ArrayList<RpcType>();
-                    message.supportedMethods.add(RpcType.valueOf(input.readEnum()));
-                    break;
-                case 9:
-                    message.encrypted = input.readBool();
-                    break;
-                case 10:
-                    message.maxWrappedSize = input.readInt32();
-                    break;
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, BitToUserHandshake message) throws IOException
-    {
-        if(message.rpcVersion != 0)
-            output.writeInt32(2, message.rpcVersion, false);
-
-        if(message.status != null)
-             output.writeEnum(3, message.status.number, false);
-
-        if(message.errorId != null)
-            output.writeString(4, message.errorId, false);
-
-        if(message.errorMessage != null)
-            output.writeString(5, message.errorMessage, false);
-
-        if(message.serverInfos != null)
-             output.writeObject(6, message.serverInfos, RpcEndpointInfos.getSchema(), false);
-
-
-        if(message.authenticationMechanisms != null)
-        {
-            for(String authenticationMechanisms : message.authenticationMechanisms)
-            {
-                if(authenticationMechanisms != null)
-                    output.writeString(7, authenticationMechanisms, true);
-            }
-        }
-
-        if(message.supportedMethods != null)
-        {
-            for(RpcType supportedMethods : message.supportedMethods)
-            {
-                if(supportedMethods != null)
-                    output.writeEnum(8, supportedMethods.number, true);
-            }
-        }
-
-        if(message.encrypted != null)
-            output.writeBool(9, message.encrypted, false);
-
-        if(message.maxWrappedSize != 0)
-            output.writeInt32(10, message.maxWrappedSize, false);
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 2: return "rpcVersion";
-            case 3: return "status";
-            case 4: return "errorId";
-            case 5: return "errorMessage";
-            case 6: return "serverInfos";
-            case 7: return "authenticationMechanisms";
-            case 8: return "supportedMethods";
-            case 9: return "encrypted";
-            case 10: return "maxWrappedSize";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("rpcVersion", 2);
-        __fieldMap.put("status", 3);
-        __fieldMap.put("errorId", 4);
-        __fieldMap.put("errorMessage", 5);
-        __fieldMap.put("serverInfos", 6);
-        __fieldMap.put("authenticationMechanisms", 7);
-        __fieldMap.put("supportedMethods", 8);
-        __fieldMap.put("encrypted", 9);
-        __fieldMap.put("maxWrappedSize", 10);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/CatalogMetadata.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/CatalogMetadata.java
deleted file mode 100644
index ea4f946..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/CatalogMetadata.java
+++ /dev/null
@@ -1,207 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class CatalogMetadata implements Externalizable, Message<CatalogMetadata>, Schema<CatalogMetadata>
-{
-
-    public static Schema<CatalogMetadata> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static CatalogMetadata getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final CatalogMetadata DEFAULT_INSTANCE = new CatalogMetadata();
-
-    
-    private String catalogName;
-    private String description;
-    private String connect;
-
-    public CatalogMetadata()
-    {
-        
-    }
-
-    // getters and setters
-
-    // catalogName
-
-    public String getCatalogName()
-    {
-        return catalogName;
-    }
-
-    public CatalogMetadata setCatalogName(String catalogName)
-    {
-        this.catalogName = catalogName;
-        return this;
-    }
-
-    // description
-
-    public String getDescription()
-    {
-        return description;
-    }
-
-    public CatalogMetadata setDescription(String description)
-    {
-        this.description = description;
-        return this;
-    }
-
-    // connect
-
-    public String getConnect()
-    {
-        return connect;
-    }
-
-    public CatalogMetadata setConnect(String connect)
-    {
-        this.connect = connect;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<CatalogMetadata> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public CatalogMetadata newMessage()
-    {
-        return new CatalogMetadata();
-    }
-
-    public Class<CatalogMetadata> typeClass()
-    {
-        return CatalogMetadata.class;
-    }
-
-    public String messageName()
-    {
-        return CatalogMetadata.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return CatalogMetadata.class.getName();
-    }
-
-    public boolean isInitialized(CatalogMetadata message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, CatalogMetadata message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.catalogName = input.readString();
-                    break;
-                case 2:
-                    message.description = input.readString();
-                    break;
-                case 3:
-                    message.connect = input.readString();
-                    break;
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, CatalogMetadata message) throws IOException
-    {
-        if(message.catalogName != null)
-            output.writeString(1, message.catalogName, false);
-
-        if(message.description != null)
-            output.writeString(2, message.description, false);
-
-        if(message.connect != null)
-            output.writeString(3, message.connect, false);
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "catalogName";
-            case 2: return "description";
-            case 3: return "connect";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("catalogName", 1);
-        __fieldMap.put("description", 2);
-        __fieldMap.put("connect", 3);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/CollateSupport.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/CollateSupport.java
deleted file mode 100644
index bc51806..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/CollateSupport.java
+++ /dev/null
@@ -1,49 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-public enum CollateSupport implements com.dyuproject.protostuff.EnumLite<CollateSupport>
-{
-    CS_UNKNOWN(0),
-    CS_GROUP_BY(1);
-    
-    public final int number;
-    
-    private CollateSupport (int number)
-    {
-        this.number = number;
-    }
-    
-    public int getNumber()
-    {
-        return number;
-    }
-    
-    public static CollateSupport valueOf(int number)
-    {
-        switch(number) 
-        {
-            case 0: return CS_UNKNOWN;
-            case 1: return CS_GROUP_BY;
-            default: return null;
-        }
-    }
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/Collector.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/Collector.java
deleted file mode 100644
index 1cce58c..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/Collector.java
+++ /dev/null
@@ -1,239 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-import java.util.ArrayList;
-import java.util.List;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class Collector implements Externalizable, Message<Collector>, Schema<Collector>
-{
-
-    public static Schema<Collector> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static Collector getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final Collector DEFAULT_INSTANCE = new Collector();
-
-    
-    private int oppositeMajorFragmentId;
-    private List<Integer> incomingMinorFragment;
-    private Boolean supportsOutOfOrder;
-    private Boolean isSpooling;
-
-    public Collector()
-    {
-        
-    }
-
-    // getters and setters
-
-    // oppositeMajorFragmentId
-
-    public int getOppositeMajorFragmentId()
-    {
-        return oppositeMajorFragmentId;
-    }
-
-    public Collector setOppositeMajorFragmentId(int oppositeMajorFragmentId)
-    {
-        this.oppositeMajorFragmentId = oppositeMajorFragmentId;
-        return this;
-    }
-
-    // incomingMinorFragment
-
-    public List<Integer> getIncomingMinorFragmentList()
-    {
-        return incomingMinorFragment;
-    }
-
-    public Collector setIncomingMinorFragmentList(List<Integer> incomingMinorFragment)
-    {
-        this.incomingMinorFragment = incomingMinorFragment;
-        return this;
-    }
-
-    // supportsOutOfOrder
-
-    public Boolean getSupportsOutOfOrder()
-    {
-        return supportsOutOfOrder;
-    }
-
-    public Collector setSupportsOutOfOrder(Boolean supportsOutOfOrder)
-    {
-        this.supportsOutOfOrder = supportsOutOfOrder;
-        return this;
-    }
-
-    // isSpooling
-
-    public Boolean getIsSpooling()
-    {
-        return isSpooling;
-    }
-
-    public Collector setIsSpooling(Boolean isSpooling)
-    {
-        this.isSpooling = isSpooling;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<Collector> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public Collector newMessage()
-    {
-        return new Collector();
-    }
-
-    public Class<Collector> typeClass()
-    {
-        return Collector.class;
-    }
-
-    public String messageName()
-    {
-        return Collector.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return Collector.class.getName();
-    }
-
-    public boolean isInitialized(Collector message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, Collector message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.oppositeMajorFragmentId = input.readInt32();
-                    break;
-                case 2:
-                    if(message.incomingMinorFragment == null)
-                        message.incomingMinorFragment = new ArrayList<Integer>();
-                    message.incomingMinorFragment.add(input.readInt32());
-                    break;
-                case 3:
-                    message.supportsOutOfOrder = input.readBool();
-                    break;
-                case 4:
-                    message.isSpooling = input.readBool();
-                    break;
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, Collector message) throws IOException
-    {
-        if(message.oppositeMajorFragmentId != 0)
-            output.writeInt32(1, message.oppositeMajorFragmentId, false);
-
-        if(message.incomingMinorFragment != null)
-        {
-            for(Integer incomingMinorFragment : message.incomingMinorFragment)
-            {
-                if(incomingMinorFragment != null)
-                    output.writeInt32(2, incomingMinorFragment, true);
-            }
-        }
-
-        if(message.supportsOutOfOrder != null)
-            output.writeBool(3, message.supportsOutOfOrder, false);
-
-        if(message.isSpooling != null)
-            output.writeBool(4, message.isSpooling, false);
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "oppositeMajorFragmentId";
-            case 2: return "incomingMinorFragment";
-            case 3: return "supportsOutOfOrder";
-            case 4: return "isSpooling";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("oppositeMajorFragmentId", 1);
-        __fieldMap.put("incomingMinorFragment", 2);
-        __fieldMap.put("supportsOutOfOrder", 3);
-        __fieldMap.put("isSpooling", 4);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/ColumnMetadata.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/ColumnMetadata.java
deleted file mode 100644
index 86eb2c6..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/ColumnMetadata.java
+++ /dev/null
@@ -1,515 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class ColumnMetadata implements Externalizable, Message<ColumnMetadata>, Schema<ColumnMetadata>
-{
-
-    public static Schema<ColumnMetadata> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static ColumnMetadata getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final ColumnMetadata DEFAULT_INSTANCE = new ColumnMetadata();
-
-    
-    private String catalogName;
-    private String schemaName;
-    private String tableName;
-    private String columnName;
-    private int ordinalPosition;
-    private String defaultValue;
-    private Boolean isNullable;
-    private String dataType;
-    private int charMaxLength;
-    private int charOctetLength;
-    private int numericPrecision;
-    private int numericPrecisionRadix;
-    private int numericScale;
-    private int dateTimePrecision;
-    private String intervalType;
-    private int intervalPrecision;
-    private int columnSize;
-
-    public ColumnMetadata()
-    {
-        
-    }
-
-    // getters and setters
-
-    // catalogName
-
-    public String getCatalogName()
-    {
-        return catalogName;
-    }
-
-    public ColumnMetadata setCatalogName(String catalogName)
-    {
-        this.catalogName = catalogName;
-        return this;
-    }
-
-    // schemaName
-
-    public String getSchemaName()
-    {
-        return schemaName;
-    }
-
-    public ColumnMetadata setSchemaName(String schemaName)
-    {
-        this.schemaName = schemaName;
-        return this;
-    }
-
-    // tableName
-
-    public String getTableName()
-    {
-        return tableName;
-    }
-
-    public ColumnMetadata setTableName(String tableName)
-    {
-        this.tableName = tableName;
-        return this;
-    }
-
-    // columnName
-
-    public String getColumnName()
-    {
-        return columnName;
-    }
-
-    public ColumnMetadata setColumnName(String columnName)
-    {
-        this.columnName = columnName;
-        return this;
-    }
-
-    // ordinalPosition
-
-    public int getOrdinalPosition()
-    {
-        return ordinalPosition;
-    }
-
-    public ColumnMetadata setOrdinalPosition(int ordinalPosition)
-    {
-        this.ordinalPosition = ordinalPosition;
-        return this;
-    }
-
-    // defaultValue
-
-    public String getDefaultValue()
-    {
-        return defaultValue;
-    }
-
-    public ColumnMetadata setDefaultValue(String defaultValue)
-    {
-        this.defaultValue = defaultValue;
-        return this;
-    }
-
-    // isNullable
-
-    public Boolean getIsNullable()
-    {
-        return isNullable;
-    }
-
-    public ColumnMetadata setIsNullable(Boolean isNullable)
-    {
-        this.isNullable = isNullable;
-        return this;
-    }
-
-    // dataType
-
-    public String getDataType()
-    {
-        return dataType;
-    }
-
-    public ColumnMetadata setDataType(String dataType)
-    {
-        this.dataType = dataType;
-        return this;
-    }
-
-    // charMaxLength
-
-    public int getCharMaxLength()
-    {
-        return charMaxLength;
-    }
-
-    public ColumnMetadata setCharMaxLength(int charMaxLength)
-    {
-        this.charMaxLength = charMaxLength;
-        return this;
-    }
-
-    // charOctetLength
-
-    public int getCharOctetLength()
-    {
-        return charOctetLength;
-    }
-
-    public ColumnMetadata setCharOctetLength(int charOctetLength)
-    {
-        this.charOctetLength = charOctetLength;
-        return this;
-    }
-
-    // numericPrecision
-
-    public int getNumericPrecision()
-    {
-        return numericPrecision;
-    }
-
-    public ColumnMetadata setNumericPrecision(int numericPrecision)
-    {
-        this.numericPrecision = numericPrecision;
-        return this;
-    }
-
-    // numericPrecisionRadix
-
-    public int getNumericPrecisionRadix()
-    {
-        return numericPrecisionRadix;
-    }
-
-    public ColumnMetadata setNumericPrecisionRadix(int numericPrecisionRadix)
-    {
-        this.numericPrecisionRadix = numericPrecisionRadix;
-        return this;
-    }
-
-    // numericScale
-
-    public int getNumericScale()
-    {
-        return numericScale;
-    }
-
-    public ColumnMetadata setNumericScale(int numericScale)
-    {
-        this.numericScale = numericScale;
-        return this;
-    }
-
-    // dateTimePrecision
-
-    public int getDateTimePrecision()
-    {
-        return dateTimePrecision;
-    }
-
-    public ColumnMetadata setDateTimePrecision(int dateTimePrecision)
-    {
-        this.dateTimePrecision = dateTimePrecision;
-        return this;
-    }
-
-    // intervalType
-
-    public String getIntervalType()
-    {
-        return intervalType;
-    }
-
-    public ColumnMetadata setIntervalType(String intervalType)
-    {
-        this.intervalType = intervalType;
-        return this;
-    }
-
-    // intervalPrecision
-
-    public int getIntervalPrecision()
-    {
-        return intervalPrecision;
-    }
-
-    public ColumnMetadata setIntervalPrecision(int intervalPrecision)
-    {
-        this.intervalPrecision = intervalPrecision;
-        return this;
-    }
-
-    // columnSize
-
-    public int getColumnSize()
-    {
-        return columnSize;
-    }
-
-    public ColumnMetadata setColumnSize(int columnSize)
-    {
-        this.columnSize = columnSize;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<ColumnMetadata> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public ColumnMetadata newMessage()
-    {
-        return new ColumnMetadata();
-    }
-
-    public Class<ColumnMetadata> typeClass()
-    {
-        return ColumnMetadata.class;
-    }
-
-    public String messageName()
-    {
-        return ColumnMetadata.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return ColumnMetadata.class.getName();
-    }
-
-    public boolean isInitialized(ColumnMetadata message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, ColumnMetadata message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.catalogName = input.readString();
-                    break;
-                case 2:
-                    message.schemaName = input.readString();
-                    break;
-                case 3:
-                    message.tableName = input.readString();
-                    break;
-                case 4:
-                    message.columnName = input.readString();
-                    break;
-                case 5:
-                    message.ordinalPosition = input.readInt32();
-                    break;
-                case 6:
-                    message.defaultValue = input.readString();
-                    break;
-                case 7:
-                    message.isNullable = input.readBool();
-                    break;
-                case 8:
-                    message.dataType = input.readString();
-                    break;
-                case 9:
-                    message.charMaxLength = input.readInt32();
-                    break;
-                case 10:
-                    message.charOctetLength = input.readInt32();
-                    break;
-                case 11:
-                    message.numericPrecision = input.readInt32();
-                    break;
-                case 12:
-                    message.numericPrecisionRadix = input.readInt32();
-                    break;
-                case 13:
-                    message.numericScale = input.readInt32();
-                    break;
-                case 14:
-                    message.dateTimePrecision = input.readInt32();
-                    break;
-                case 15:
-                    message.intervalType = input.readString();
-                    break;
-                case 16:
-                    message.intervalPrecision = input.readInt32();
-                    break;
-                case 17:
-                    message.columnSize = input.readInt32();
-                    break;
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, ColumnMetadata message) throws IOException
-    {
-        if(message.catalogName != null)
-            output.writeString(1, message.catalogName, false);
-
-        if(message.schemaName != null)
-            output.writeString(2, message.schemaName, false);
-
-        if(message.tableName != null)
-            output.writeString(3, message.tableName, false);
-
-        if(message.columnName != null)
-            output.writeString(4, message.columnName, false);
-
-        if(message.ordinalPosition != 0)
-            output.writeInt32(5, message.ordinalPosition, false);
-
-        if(message.defaultValue != null)
-            output.writeString(6, message.defaultValue, false);
-
-        if(message.isNullable != null)
-            output.writeBool(7, message.isNullable, false);
-
-        if(message.dataType != null)
-            output.writeString(8, message.dataType, false);
-
-        if(message.charMaxLength != 0)
-            output.writeInt32(9, message.charMaxLength, false);
-
-        if(message.charOctetLength != 0)
-            output.writeInt32(10, message.charOctetLength, false);
-
-        if(message.numericPrecision != 0)
-            output.writeInt32(11, message.numericPrecision, false);
-
-        if(message.numericPrecisionRadix != 0)
-            output.writeInt32(12, message.numericPrecisionRadix, false);
-
-        if(message.numericScale != 0)
-            output.writeInt32(13, message.numericScale, false);
-
-        if(message.dateTimePrecision != 0)
-            output.writeInt32(14, message.dateTimePrecision, false);
-
-        if(message.intervalType != null)
-            output.writeString(15, message.intervalType, false);
-
-        if(message.intervalPrecision != 0)
-            output.writeInt32(16, message.intervalPrecision, false);
-
-        if(message.columnSize != 0)
-            output.writeInt32(17, message.columnSize, false);
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "catalogName";
-            case 2: return "schemaName";
-            case 3: return "tableName";
-            case 4: return "columnName";
-            case 5: return "ordinalPosition";
-            case 6: return "defaultValue";
-            case 7: return "isNullable";
-            case 8: return "dataType";
-            case 9: return "charMaxLength";
-            case 10: return "charOctetLength";
-            case 11: return "numericPrecision";
-            case 12: return "numericPrecisionRadix";
-            case 13: return "numericScale";
-            case 14: return "dateTimePrecision";
-            case 15: return "intervalType";
-            case 16: return "intervalPrecision";
-            case 17: return "columnSize";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("catalogName", 1);
-        __fieldMap.put("schemaName", 2);
-        __fieldMap.put("tableName", 3);
-        __fieldMap.put("columnName", 4);
-        __fieldMap.put("ordinalPosition", 5);
-        __fieldMap.put("defaultValue", 6);
-        __fieldMap.put("isNullable", 7);
-        __fieldMap.put("dataType", 8);
-        __fieldMap.put("charMaxLength", 9);
-        __fieldMap.put("charOctetLength", 10);
-        __fieldMap.put("numericPrecision", 11);
-        __fieldMap.put("numericPrecisionRadix", 12);
-        __fieldMap.put("numericScale", 13);
-        __fieldMap.put("dateTimePrecision", 14);
-        __fieldMap.put("intervalType", 15);
-        __fieldMap.put("intervalPrecision", 16);
-        __fieldMap.put("columnSize", 17);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/ColumnSearchability.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/ColumnSearchability.java
deleted file mode 100644
index 508fe88..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/ColumnSearchability.java
+++ /dev/null
@@ -1,55 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-public enum ColumnSearchability implements com.dyuproject.protostuff.EnumLite<ColumnSearchability>
-{
-    UNKNOWN_SEARCHABILITY(0),
-    NONE(1),
-    CHAR(2),
-    NUMBER(3),
-    ALL(4);
-    
-    public final int number;
-    
-    private ColumnSearchability (int number)
-    {
-        this.number = number;
-    }
-    
-    public int getNumber()
-    {
-        return number;
-    }
-    
-    public static ColumnSearchability valueOf(int number)
-    {
-        switch(number) 
-        {
-            case 0: return UNKNOWN_SEARCHABILITY;
-            case 1: return NONE;
-            case 2: return CHAR;
-            case 3: return NUMBER;
-            case 4: return ALL;
-            default: return null;
-        }
-    }
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/ColumnUpdatability.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/ColumnUpdatability.java
deleted file mode 100644
index 9470adb..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/ColumnUpdatability.java
+++ /dev/null
@@ -1,51 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-public enum ColumnUpdatability implements com.dyuproject.protostuff.EnumLite<ColumnUpdatability>
-{
-    UNKNOWN_UPDATABILITY(0),
-    READ_ONLY(1),
-    WRITABLE(2);
-    
-    public final int number;
-    
-    private ColumnUpdatability (int number)
-    {
-        this.number = number;
-    }
-    
-    public int getNumber()
-    {
-        return number;
-    }
-    
-    public static ColumnUpdatability valueOf(int number)
-    {
-        switch(number) 
-        {
-            case 0: return UNKNOWN_UPDATABILITY;
-            case 1: return READ_ONLY;
-            case 2: return WRITABLE;
-            default: return null;
-        }
-    }
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/CompleteRpcMessage.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/CompleteRpcMessage.java
deleted file mode 100644
index f6d578c..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/CompleteRpcMessage.java
+++ /dev/null
@@ -1,210 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-
-import com.dyuproject.protostuff.ByteString;
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class CompleteRpcMessage implements Externalizable, Message<CompleteRpcMessage>, Schema<CompleteRpcMessage>
-{
-
-    public static Schema<CompleteRpcMessage> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static CompleteRpcMessage getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final CompleteRpcMessage DEFAULT_INSTANCE = new CompleteRpcMessage();
-
-    
-    private RpcHeader header;
-    private ByteString protobufBody;
-    private ByteString rawBody;
-
-    public CompleteRpcMessage()
-    {
-        
-    }
-
-    // getters and setters
-
-    // header
-
-    public RpcHeader getHeader()
-    {
-        return header;
-    }
-
-    public CompleteRpcMessage setHeader(RpcHeader header)
-    {
-        this.header = header;
-        return this;
-    }
-
-    // protobufBody
-
-    public ByteString getProtobufBody()
-    {
-        return protobufBody;
-    }
-
-    public CompleteRpcMessage setProtobufBody(ByteString protobufBody)
-    {
-        this.protobufBody = protobufBody;
-        return this;
-    }
-
-    // rawBody
-
-    public ByteString getRawBody()
-    {
-        return rawBody;
-    }
-
-    public CompleteRpcMessage setRawBody(ByteString rawBody)
-    {
-        this.rawBody = rawBody;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<CompleteRpcMessage> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public CompleteRpcMessage newMessage()
-    {
-        return new CompleteRpcMessage();
-    }
-
-    public Class<CompleteRpcMessage> typeClass()
-    {
-        return CompleteRpcMessage.class;
-    }
-
-    public String messageName()
-    {
-        return CompleteRpcMessage.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return CompleteRpcMessage.class.getName();
-    }
-
-    public boolean isInitialized(CompleteRpcMessage message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, CompleteRpcMessage message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.header = input.mergeObject(message.header, RpcHeader.getSchema());
-                    break;
-
-                case 2:
-                    message.protobufBody = input.readBytes();
-                    break;
-                case 3:
-                    message.rawBody = input.readBytes();
-                    break;
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, CompleteRpcMessage message) throws IOException
-    {
-        if(message.header != null)
-             output.writeObject(1, message.header, RpcHeader.getSchema(), false);
-
-
-        if(message.protobufBody != null)
-            output.writeBytes(2, message.protobufBody, false);
-
-        if(message.rawBody != null)
-            output.writeBytes(3, message.rawBody, false);
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "header";
-            case 2: return "protobufBody";
-            case 3: return "rawBody";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("header", 1);
-        __fieldMap.put("protobufBody", 2);
-        __fieldMap.put("rawBody", 3);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/ConvertSupport.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/ConvertSupport.java
deleted file mode 100644
index 96da4c1..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/ConvertSupport.java
+++ /dev/null
@@ -1,199 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-import com.dyuproject.protostuff.UninitializedMessageException;
-
-public final class ConvertSupport implements Externalizable, Message<ConvertSupport>, Schema<ConvertSupport>
-{
-
-    public static Schema<ConvertSupport> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static ConvertSupport getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final ConvertSupport DEFAULT_INSTANCE = new ConvertSupport();
-
-    
-    private org.apache.drill.common.types.MinorType from;
-    private org.apache.drill.common.types.MinorType to;
-
-    public ConvertSupport()
-    {
-        
-    }
-
-    public ConvertSupport(
-        org.apache.drill.common.types.MinorType from,
-        org.apache.drill.common.types.MinorType to
-    )
-    {
-        this.from = from;
-        this.to = to;
-    }
-
-    // getters and setters
-
-    // from
-
-    public org.apache.drill.common.types.MinorType getFrom()
-    {
-        return from;
-    }
-
-    public ConvertSupport setFrom(org.apache.drill.common.types.MinorType from)
-    {
-        this.from = from;
-        return this;
-    }
-
-    // to
-
-    public org.apache.drill.common.types.MinorType getTo()
-    {
-        return to;
-    }
-
-    public ConvertSupport setTo(org.apache.drill.common.types.MinorType to)
-    {
-        this.to = to;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<ConvertSupport> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public ConvertSupport newMessage()
-    {
-        return new ConvertSupport();
-    }
-
-    public Class<ConvertSupport> typeClass()
-    {
-        return ConvertSupport.class;
-    }
-
-    public String messageName()
-    {
-        return ConvertSupport.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return ConvertSupport.class.getName();
-    }
-
-    public boolean isInitialized(ConvertSupport message)
-    {
-        return 
-            message.from != null 
-            && message.to != null;
-    }
-
-    public void mergeFrom(Input input, ConvertSupport message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.from = org.apache.drill.common.types.MinorType.valueOf(input.readEnum());
-                    break;
-                case 2:
-                    message.to = org.apache.drill.common.types.MinorType.valueOf(input.readEnum());
-                    break;
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, ConvertSupport message) throws IOException
-    {
-        if(message.from == null)
-            throw new UninitializedMessageException(message);
-        output.writeEnum(1, message.from.number, false);
-
-        if(message.to == null)
-            throw new UninitializedMessageException(message);
-        output.writeEnum(2, message.to.number, false);
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "from";
-            case 2: return "to";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("from", 1);
-        __fieldMap.put("to", 2);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/CoreOperatorType.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/CoreOperatorType.java
deleted file mode 100644
index eea96ca..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/CoreOperatorType.java
+++ /dev/null
@@ -1,171 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-public enum CoreOperatorType implements com.dyuproject.protostuff.EnumLite<CoreOperatorType>
-{
-    SINGLE_SENDER(0),
-    BROADCAST_SENDER(1),
-    FILTER(2),
-    HASH_AGGREGATE(3),
-    HASH_JOIN(4),
-    MERGE_JOIN(5),
-    HASH_PARTITION_SENDER(6),
-    LIMIT(7),
-    MERGING_RECEIVER(8),
-    ORDERED_PARTITION_SENDER(9),
-    PROJECT(10),
-    UNORDERED_RECEIVER(11),
-    RANGE_PARTITION_SENDER(12),
-    SCREEN(13),
-    SELECTION_VECTOR_REMOVER(14),
-    STREAMING_AGGREGATE(15),
-    TOP_N_SORT(16),
-    EXTERNAL_SORT(17),
-    TRACE(18),
-    UNION(19),
-    OLD_SORT(20),
-    PARQUET_ROW_GROUP_SCAN(21),
-    HIVE_SUB_SCAN(22),
-    SYSTEM_TABLE_SCAN(23),
-    MOCK_SUB_SCAN(24),
-    PARQUET_WRITER(25),
-    DIRECT_SUB_SCAN(26),
-    TEXT_WRITER(27),
-    TEXT_SUB_SCAN(28),
-    JSON_SUB_SCAN(29),
-    INFO_SCHEMA_SUB_SCAN(30),
-    COMPLEX_TO_JSON(31),
-    PRODUCER_CONSUMER(32),
-    HBASE_SUB_SCAN(33),
-    WINDOW(34),
-    NESTED_LOOP_JOIN(35),
-    AVRO_SUB_SCAN(36),
-    PCAP_SUB_SCAN(37),
-    KAFKA_SUB_SCAN(38),
-    KUDU_SUB_SCAN(39),
-    FLATTEN(40),
-    LATERAL_JOIN(41),
-    UNNEST(42),
-    HIVE_DRILL_NATIVE_PARQUET_ROW_GROUP_SCAN(43),
-    JDBC_SCAN(44),
-    REGEX_SUB_SCAN(45),
-    MAPRDB_SUB_SCAN(46),
-    MONGO_SUB_SCAN(47),
-    KUDU_WRITER(48),
-    OPEN_TSDB_SUB_SCAN(49),
-    JSON_WRITER(50),
-    HTPPD_LOG_SUB_SCAN(51),
-    IMAGE_SUB_SCAN(52),
-    SEQUENCE_SUB_SCAN(53),
-    PARTITION_LIMIT(54),
-    PCAPNG_SUB_SCAN(55),
-    RUNTIME_FILTER(56),
-    ROWKEY_JOIN(57),
-    SYSLOG_SUB_SCAN(58),
-    STATISTICS_AGGREGATE(59),
-    UNPIVOT_MAPS(60),
-    STATISTICS_MERGE(61),
-    LTSV_SUB_SCAN(62);
-    
-    public final int number;
-    
-    private CoreOperatorType (int number)
-    {
-        this.number = number;
-    }
-    
-    public int getNumber()
-    {
-        return number;
-    }
-    
-    public static CoreOperatorType valueOf(int number)
-    {
-        switch(number) 
-        {
-            case 0: return SINGLE_SENDER;
-            case 1: return BROADCAST_SENDER;
-            case 2: return FILTER;
-            case 3: return HASH_AGGREGATE;
-            case 4: return HASH_JOIN;
-            case 5: return MERGE_JOIN;
-            case 6: return HASH_PARTITION_SENDER;
-            case 7: return LIMIT;
-            case 8: return MERGING_RECEIVER;
-            case 9: return ORDERED_PARTITION_SENDER;
-            case 10: return PROJECT;
-            case 11: return UNORDERED_RECEIVER;
-            case 12: return RANGE_PARTITION_SENDER;
-            case 13: return SCREEN;
-            case 14: return SELECTION_VECTOR_REMOVER;
-            case 15: return STREAMING_AGGREGATE;
-            case 16: return TOP_N_SORT;
-            case 17: return EXTERNAL_SORT;
-            case 18: return TRACE;
-            case 19: return UNION;
-            case 20: return OLD_SORT;
-            case 21: return PARQUET_ROW_GROUP_SCAN;
-            case 22: return HIVE_SUB_SCAN;
-            case 23: return SYSTEM_TABLE_SCAN;
-            case 24: return MOCK_SUB_SCAN;
-            case 25: return PARQUET_WRITER;
-            case 26: return DIRECT_SUB_SCAN;
-            case 27: return TEXT_WRITER;
-            case 28: return TEXT_SUB_SCAN;
-            case 29: return JSON_SUB_SCAN;
-            case 30: return INFO_SCHEMA_SUB_SCAN;
-            case 31: return COMPLEX_TO_JSON;
-            case 32: return PRODUCER_CONSUMER;
-            case 33: return HBASE_SUB_SCAN;
-            case 34: return WINDOW;
-            case 35: return NESTED_LOOP_JOIN;
-            case 36: return AVRO_SUB_SCAN;
-            case 37: return PCAP_SUB_SCAN;
-            case 38: return KAFKA_SUB_SCAN;
-            case 39: return KUDU_SUB_SCAN;
-            case 40: return FLATTEN;
-            case 41: return LATERAL_JOIN;
-            case 42: return UNNEST;
-            case 43: return HIVE_DRILL_NATIVE_PARQUET_ROW_GROUP_SCAN;
-            case 44: return JDBC_SCAN;
-            case 45: return REGEX_SUB_SCAN;
-            case 46: return MAPRDB_SUB_SCAN;
-            case 47: return MONGO_SUB_SCAN;
-            case 48: return KUDU_WRITER;
-            case 49: return OPEN_TSDB_SUB_SCAN;
-            case 50: return JSON_WRITER;
-            case 51: return HTPPD_LOG_SUB_SCAN;
-            case 52: return IMAGE_SUB_SCAN;
-            case 53: return SEQUENCE_SUB_SCAN;
-            case 54: return PARTITION_LIMIT;
-            case 55: return PCAPNG_SUB_SCAN;
-            case 56: return RUNTIME_FILTER;
-            case 57: return ROWKEY_JOIN;
-            case 58: return SYSLOG_SUB_SCAN;
-            case 59: return STATISTICS_AGGREGATE;
-            case 60: return UNPIVOT_MAPS;
-            case 61: return STATISTICS_MERGE;
-            case 62: return LTSV_SUB_SCAN;
-            default: return null;
-        }
-    }
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/CorrelationNamesSupport.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/CorrelationNamesSupport.java
deleted file mode 100644
index 061ea6d..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/CorrelationNamesSupport.java
+++ /dev/null
@@ -1,51 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-public enum CorrelationNamesSupport implements com.dyuproject.protostuff.EnumLite<CorrelationNamesSupport>
-{
-    CN_NONE(1),
-    CN_DIFFERENT_NAMES(2),
-    CN_ANY(3);
-    
-    public final int number;
-    
-    private CorrelationNamesSupport (int number)
-    {
-        this.number = number;
-    }
-    
-    public int getNumber()
-    {
-        return number;
-    }
-    
-    public static CorrelationNamesSupport valueOf(int number)
-    {
-        switch(number) 
-        {
-            case 1: return CN_NONE;
-            case 2: return CN_DIFFERENT_NAMES;
-            case 3: return CN_ANY;
-            default: return null;
-        }
-    }
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/CreatePreparedStatementReq.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/CreatePreparedStatementReq.java
deleted file mode 100644
index c21015f..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/CreatePreparedStatementReq.java
+++ /dev/null
@@ -1,163 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class CreatePreparedStatementReq implements Externalizable, Message<CreatePreparedStatementReq>, Schema<CreatePreparedStatementReq>
-{
-
-    public static Schema<CreatePreparedStatementReq> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static CreatePreparedStatementReq getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final CreatePreparedStatementReq DEFAULT_INSTANCE = new CreatePreparedStatementReq();
-
-    
-    private String sqlQuery;
-
-    public CreatePreparedStatementReq()
-    {
-        
-    }
-
-    // getters and setters
-
-    // sqlQuery
-
-    public String getSqlQuery()
-    {
-        return sqlQuery;
-    }
-
-    public CreatePreparedStatementReq setSqlQuery(String sqlQuery)
-    {
-        this.sqlQuery = sqlQuery;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<CreatePreparedStatementReq> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public CreatePreparedStatementReq newMessage()
-    {
-        return new CreatePreparedStatementReq();
-    }
-
-    public Class<CreatePreparedStatementReq> typeClass()
-    {
-        return CreatePreparedStatementReq.class;
-    }
-
-    public String messageName()
-    {
-        return CreatePreparedStatementReq.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return CreatePreparedStatementReq.class.getName();
-    }
-
-    public boolean isInitialized(CreatePreparedStatementReq message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, CreatePreparedStatementReq message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.sqlQuery = input.readString();
-                    break;
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, CreatePreparedStatementReq message) throws IOException
-    {
-        if(message.sqlQuery != null)
-            output.writeString(1, message.sqlQuery, false);
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "sqlQuery";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("sqlQuery", 1);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/CreatePreparedStatementResp.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/CreatePreparedStatementResp.java
deleted file mode 100644
index 452af40..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/CreatePreparedStatementResp.java
+++ /dev/null
@@ -1,211 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class CreatePreparedStatementResp implements Externalizable, Message<CreatePreparedStatementResp>, Schema<CreatePreparedStatementResp>
-{
-
-    public static Schema<CreatePreparedStatementResp> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static CreatePreparedStatementResp getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final CreatePreparedStatementResp DEFAULT_INSTANCE = new CreatePreparedStatementResp();
-
-    
-    private RequestStatus status;
-    private PreparedStatement preparedStatement;
-    private DrillPBError error;
-
-    public CreatePreparedStatementResp()
-    {
-        
-    }
-
-    // getters and setters
-
-    // status
-
-    public RequestStatus getStatus()
-    {
-        return status == null ? RequestStatus.UNKNOWN_STATUS : status;
-    }
-
-    public CreatePreparedStatementResp setStatus(RequestStatus status)
-    {
-        this.status = status;
-        return this;
-    }
-
-    // preparedStatement
-
-    public PreparedStatement getPreparedStatement()
-    {
-        return preparedStatement;
-    }
-
-    public CreatePreparedStatementResp setPreparedStatement(PreparedStatement preparedStatement)
-    {
-        this.preparedStatement = preparedStatement;
-        return this;
-    }
-
-    // error
-
-    public DrillPBError getError()
-    {
-        return error;
-    }
-
-    public CreatePreparedStatementResp setError(DrillPBError error)
-    {
-        this.error = error;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<CreatePreparedStatementResp> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public CreatePreparedStatementResp newMessage()
-    {
-        return new CreatePreparedStatementResp();
-    }
-
-    public Class<CreatePreparedStatementResp> typeClass()
-    {
-        return CreatePreparedStatementResp.class;
-    }
-
-    public String messageName()
-    {
-        return CreatePreparedStatementResp.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return CreatePreparedStatementResp.class.getName();
-    }
-
-    public boolean isInitialized(CreatePreparedStatementResp message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, CreatePreparedStatementResp message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.status = RequestStatus.valueOf(input.readEnum());
-                    break;
-                case 2:
-                    message.preparedStatement = input.mergeObject(message.preparedStatement, PreparedStatement.getSchema());
-                    break;
-
-                case 3:
-                    message.error = input.mergeObject(message.error, DrillPBError.getSchema());
-                    break;
-
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, CreatePreparedStatementResp message) throws IOException
-    {
-        if(message.status != null)
-             output.writeEnum(1, message.status.number, false);
-
-        if(message.preparedStatement != null)
-             output.writeObject(2, message.preparedStatement, PreparedStatement.getSchema(), false);
-
-
-        if(message.error != null)
-             output.writeObject(3, message.error, DrillPBError.getSchema(), false);
-
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "status";
-            case 2: return "preparedStatement";
-            case 3: return "error";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("status", 1);
-        __fieldMap.put("preparedStatement", 2);
-        __fieldMap.put("error", 3);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/CustomMessage.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/CustomMessage.java
deleted file mode 100644
index 8be7718..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/CustomMessage.java
+++ /dev/null
@@ -1,186 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-
-import com.dyuproject.protostuff.ByteString;
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class CustomMessage implements Externalizable, Message<CustomMessage>, Schema<CustomMessage>
-{
-
-    public static Schema<CustomMessage> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static CustomMessage getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final CustomMessage DEFAULT_INSTANCE = new CustomMessage();
-
-    
-    private int type;
-    private ByteString message;
-
-    public CustomMessage()
-    {
-        
-    }
-
-    // getters and setters
-
-    // type
-
-    public int getType()
-    {
-        return type;
-    }
-
-    public CustomMessage setType(int type)
-    {
-        this.type = type;
-        return this;
-    }
-
-    // message
-
-    public ByteString getMessage()
-    {
-        return message;
-    }
-
-    public CustomMessage setMessage(ByteString message)
-    {
-        this.message = message;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<CustomMessage> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public CustomMessage newMessage()
-    {
-        return new CustomMessage();
-    }
-
-    public Class<CustomMessage> typeClass()
-    {
-        return CustomMessage.class;
-    }
-
-    public String messageName()
-    {
-        return CustomMessage.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return CustomMessage.class.getName();
-    }
-
-    public boolean isInitialized(CustomMessage message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, CustomMessage message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.type = input.readInt32();
-                    break;
-                case 2:
-                    message.message = input.readBytes();
-                    break;
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, CustomMessage message) throws IOException
-    {
-        if(message.type != 0)
-            output.writeInt32(1, message.type, false);
-
-        if(message.message != null)
-            output.writeBytes(2, message.message, false);
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "type";
-            case 2: return "message";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("type", 1);
-        __fieldMap.put("message", 2);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/DateTimeLiteralsSupport.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/DateTimeLiteralsSupport.java
deleted file mode 100644
index f965b33..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/DateTimeLiteralsSupport.java
+++ /dev/null
@@ -1,79 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-public enum DateTimeLiteralsSupport implements com.dyuproject.protostuff.EnumLite<DateTimeLiteralsSupport>
-{
-    DL_UNKNOWN(0),
-    DL_DATE(1),
-    DL_TIME(2),
-    DL_TIMESTAMP(3),
-    DL_INTERVAL_YEAR(4),
-    DL_INTERVAL_MONTH(5),
-    DL_INTERVAL_DAY(6),
-    DL_INTERVAL_HOUR(7),
-    DL_INTERVAL_MINUTE(8),
-    DL_INTERVAL_SECOND(9),
-    DL_INTERVAL_YEAR_TO_MONTH(10),
-    DL_INTERVAL_DAY_TO_HOUR(11),
-    DL_INTERVAL_DAY_TO_MINUTE(12),
-    DL_INTERVAL_DAY_TO_SECOND(13),
-    DL_INTERVAL_HOUR_TO_MINUTE(14),
-    DL_INTERVAL_HOUR_TO_SECOND(15),
-    DL_INTERVAL_MINUTE_TO_SECOND(16);
-    
-    public final int number;
-    
-    private DateTimeLiteralsSupport (int number)
-    {
-        this.number = number;
-    }
-    
-    public int getNumber()
-    {
-        return number;
-    }
-    
-    public static DateTimeLiteralsSupport valueOf(int number)
-    {
-        switch(number) 
-        {
-            case 0: return DL_UNKNOWN;
-            case 1: return DL_DATE;
-            case 2: return DL_TIME;
-            case 3: return DL_TIMESTAMP;
-            case 4: return DL_INTERVAL_YEAR;
-            case 5: return DL_INTERVAL_MONTH;
-            case 6: return DL_INTERVAL_DAY;
-            case 7: return DL_INTERVAL_HOUR;
-            case 8: return DL_INTERVAL_MINUTE;
-            case 9: return DL_INTERVAL_SECOND;
-            case 10: return DL_INTERVAL_YEAR_TO_MONTH;
-            case 11: return DL_INTERVAL_DAY_TO_HOUR;
-            case 12: return DL_INTERVAL_DAY_TO_MINUTE;
-            case 13: return DL_INTERVAL_DAY_TO_SECOND;
-            case 14: return DL_INTERVAL_HOUR_TO_MINUTE;
-            case 15: return DL_INTERVAL_HOUR_TO_SECOND;
-            case 16: return DL_INTERVAL_MINUTE_TO_SECOND;
-            default: return null;
-        }
-    }
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/DrillPBError.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/DrillPBError.java
deleted file mode 100644
index 172dd96..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/DrillPBError.java
+++ /dev/null
@@ -1,341 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-import java.util.ArrayList;
-import java.util.List;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class DrillPBError implements Externalizable, Message<DrillPBError>, Schema<DrillPBError>
-{
-    public enum ErrorType implements com.dyuproject.protostuff.EnumLite<ErrorType>
-    {
-        CONNECTION(0),
-        DATA_READ(1),
-        DATA_WRITE(2),
-        FUNCTION(3),
-        PARSE(4),
-        PERMISSION(5),
-        PLAN(6),
-        RESOURCE(7),
-        SYSTEM(8),
-        UNSUPPORTED_OPERATION(9),
-        VALIDATION(10),
-        EXECUTION_ERROR(11),
-        INTERNAL_ERROR(12),
-        UNSPECIFIED_ERROR(13);
-        
-        public final int number;
-        
-        private ErrorType (int number)
-        {
-            this.number = number;
-        }
-        
-        public int getNumber()
-        {
-            return number;
-        }
-        
-        public static ErrorType valueOf(int number)
-        {
-            switch(number) 
-            {
-                case 0: return CONNECTION;
-                case 1: return DATA_READ;
-                case 2: return DATA_WRITE;
-                case 3: return FUNCTION;
-                case 4: return PARSE;
-                case 5: return PERMISSION;
-                case 6: return PLAN;
-                case 7: return RESOURCE;
-                case 8: return SYSTEM;
-                case 9: return UNSUPPORTED_OPERATION;
-                case 10: return VALIDATION;
-                case 11: return EXECUTION_ERROR;
-                case 12: return INTERNAL_ERROR;
-                case 13: return UNSPECIFIED_ERROR;
-                default: return null;
-            }
-        }
-    }
-
-
-    public static Schema<DrillPBError> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static DrillPBError getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final DrillPBError DEFAULT_INSTANCE = new DrillPBError();
-
-    
-    private String errorId;
-    private DrillbitEndpoint endpoint;
-    private ErrorType errorType;
-    private String message;
-    private ExceptionWrapper exception;
-    private List<ParsingError> parsingError;
-
-    public DrillPBError()
-    {
-        
-    }
-
-    // getters and setters
-
-    // errorId
-
-    public String getErrorId()
-    {
-        return errorId;
-    }
-
-    public DrillPBError setErrorId(String errorId)
-    {
-        this.errorId = errorId;
-        return this;
-    }
-
-    // endpoint
-
-    public DrillbitEndpoint getEndpoint()
-    {
-        return endpoint;
-    }
-
-    public DrillPBError setEndpoint(DrillbitEndpoint endpoint)
-    {
-        this.endpoint = endpoint;
-        return this;
-    }
-
-    // errorType
-
-    public ErrorType getErrorType()
-    {
-        return errorType == null ? ErrorType.CONNECTION : errorType;
-    }
-
-    public DrillPBError setErrorType(ErrorType errorType)
-    {
-        this.errorType = errorType;
-        return this;
-    }
-
-    // message
-
-    public String getMessage()
-    {
-        return message;
-    }
-
-    public DrillPBError setMessage(String message)
-    {
-        this.message = message;
-        return this;
-    }
-
-    // exception
-
-    public ExceptionWrapper getException()
-    {
-        return exception;
-    }
-
-    public DrillPBError setException(ExceptionWrapper exception)
-    {
-        this.exception = exception;
-        return this;
-    }
-
-    // parsingError
-
-    public List<ParsingError> getParsingErrorList()
-    {
-        return parsingError;
-    }
-
-    public DrillPBError setParsingErrorList(List<ParsingError> parsingError)
-    {
-        this.parsingError = parsingError;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<DrillPBError> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public DrillPBError newMessage()
-    {
-        return new DrillPBError();
-    }
-
-    public Class<DrillPBError> typeClass()
-    {
-        return DrillPBError.class;
-    }
-
-    public String messageName()
-    {
-        return DrillPBError.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return DrillPBError.class.getName();
-    }
-
-    public boolean isInitialized(DrillPBError message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, DrillPBError message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.errorId = input.readString();
-                    break;
-                case 2:
-                    message.endpoint = input.mergeObject(message.endpoint, DrillbitEndpoint.getSchema());
-                    break;
-
-                case 3:
-                    message.errorType = ErrorType.valueOf(input.readEnum());
-                    break;
-                case 4:
-                    message.message = input.readString();
-                    break;
-                case 5:
-                    message.exception = input.mergeObject(message.exception, ExceptionWrapper.getSchema());
-                    break;
-
-                case 6:
-                    if(message.parsingError == null)
-                        message.parsingError = new ArrayList<ParsingError>();
-                    message.parsingError.add(input.mergeObject(null, ParsingError.getSchema()));
-                    break;
-
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, DrillPBError message) throws IOException
-    {
-        if(message.errorId != null)
-            output.writeString(1, message.errorId, false);
-
-        if(message.endpoint != null)
-             output.writeObject(2, message.endpoint, DrillbitEndpoint.getSchema(), false);
-
-
-        if(message.errorType != null)
-             output.writeEnum(3, message.errorType.number, false);
-
-        if(message.message != null)
-            output.writeString(4, message.message, false);
-
-        if(message.exception != null)
-             output.writeObject(5, message.exception, ExceptionWrapper.getSchema(), false);
-
-
-        if(message.parsingError != null)
-        {
-            for(ParsingError parsingError : message.parsingError)
-            {
-                if(parsingError != null)
-                    output.writeObject(6, parsingError, ParsingError.getSchema(), true);
-            }
-        }
-
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "errorId";
-            case 2: return "endpoint";
-            case 3: return "errorType";
-            case 4: return "message";
-            case 5: return "exception";
-            case 6: return "parsingError";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("errorId", 1);
-        __fieldMap.put("endpoint", 2);
-        __fieldMap.put("errorType", 3);
-        __fieldMap.put("message", 4);
-        __fieldMap.put("exception", 5);
-        __fieldMap.put("parsingError", 6);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/DrillServiceInstance.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/DrillServiceInstance.java
deleted file mode 100644
index 6a68b38..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/DrillServiceInstance.java
+++ /dev/null
@@ -1,209 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class DrillServiceInstance implements Externalizable, Message<DrillServiceInstance>, Schema<DrillServiceInstance>
-{
-
-    public static Schema<DrillServiceInstance> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static DrillServiceInstance getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final DrillServiceInstance DEFAULT_INSTANCE = new DrillServiceInstance();
-
-    
-    private String id;
-    private long registrationTimeUTC;
-    private DrillbitEndpoint endpoint;
-
-    public DrillServiceInstance()
-    {
-        
-    }
-
-    // getters and setters
-
-    // id
-
-    public String getId()
-    {
-        return id;
-    }
-
-    public DrillServiceInstance setId(String id)
-    {
-        this.id = id;
-        return this;
-    }
-
-    // registrationTimeUTC
-
-    public long getRegistrationTimeUTC()
-    {
-        return registrationTimeUTC;
-    }
-
-    public DrillServiceInstance setRegistrationTimeUTC(long registrationTimeUTC)
-    {
-        this.registrationTimeUTC = registrationTimeUTC;
-        return this;
-    }
-
-    // endpoint
-
-    public DrillbitEndpoint getEndpoint()
-    {
-        return endpoint;
-    }
-
-    public DrillServiceInstance setEndpoint(DrillbitEndpoint endpoint)
-    {
-        this.endpoint = endpoint;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<DrillServiceInstance> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public DrillServiceInstance newMessage()
-    {
-        return new DrillServiceInstance();
-    }
-
-    public Class<DrillServiceInstance> typeClass()
-    {
-        return DrillServiceInstance.class;
-    }
-
-    public String messageName()
-    {
-        return DrillServiceInstance.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return DrillServiceInstance.class.getName();
-    }
-
-    public boolean isInitialized(DrillServiceInstance message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, DrillServiceInstance message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.id = input.readString();
-                    break;
-                case 2:
-                    message.registrationTimeUTC = input.readInt64();
-                    break;
-                case 3:
-                    message.endpoint = input.mergeObject(message.endpoint, DrillbitEndpoint.getSchema());
-                    break;
-
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, DrillServiceInstance message) throws IOException
-    {
-        if(message.id != null)
-            output.writeString(1, message.id, false);
-
-        if(message.registrationTimeUTC != 0)
-            output.writeInt64(2, message.registrationTimeUTC, false);
-
-        if(message.endpoint != null)
-             output.writeObject(3, message.endpoint, DrillbitEndpoint.getSchema(), false);
-
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "id";
-            case 2: return "registrationTimeUTC";
-            case 3: return "endpoint";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("id", 1);
-        __fieldMap.put("registrationTimeUTC", 2);
-        __fieldMap.put("endpoint", 3);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/DrillbitEndpoint.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/DrillbitEndpoint.java
deleted file mode 100644
index 0e642c5..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/DrillbitEndpoint.java
+++ /dev/null
@@ -1,351 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class DrillbitEndpoint implements Externalizable, Message<DrillbitEndpoint>, Schema<DrillbitEndpoint>
-{
-    public enum State implements com.dyuproject.protostuff.EnumLite<State>
-    {
-        STARTUP(0),
-        ONLINE(1),
-        QUIESCENT(2),
-        OFFLINE(3);
-        
-        public final int number;
-        
-        private State (int number)
-        {
-            this.number = number;
-        }
-        
-        public int getNumber()
-        {
-            return number;
-        }
-        
-        public static State valueOf(int number)
-        {
-            switch(number) 
-            {
-                case 0: return STARTUP;
-                case 1: return ONLINE;
-                case 2: return QUIESCENT;
-                case 3: return OFFLINE;
-                default: return null;
-            }
-        }
-    }
-
-
-    public static Schema<DrillbitEndpoint> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static DrillbitEndpoint getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final DrillbitEndpoint DEFAULT_INSTANCE = new DrillbitEndpoint();
-
-    
-    private String address;
-    private int userPort;
-    private int controlPort;
-    private int dataPort;
-    private Roles roles;
-    private String version;
-    private State state;
-    private int httpPort;
-
-    public DrillbitEndpoint()
-    {
-        
-    }
-
-    // getters and setters
-
-    // address
-
-    public String getAddress()
-    {
-        return address;
-    }
-
-    public DrillbitEndpoint setAddress(String address)
-    {
-        this.address = address;
-        return this;
-    }
-
-    // userPort
-
-    public int getUserPort()
-    {
-        return userPort;
-    }
-
-    public DrillbitEndpoint setUserPort(int userPort)
-    {
-        this.userPort = userPort;
-        return this;
-    }
-
-    // controlPort
-
-    public int getControlPort()
-    {
-        return controlPort;
-    }
-
-    public DrillbitEndpoint setControlPort(int controlPort)
-    {
-        this.controlPort = controlPort;
-        return this;
-    }
-
-    // dataPort
-
-    public int getDataPort()
-    {
-        return dataPort;
-    }
-
-    public DrillbitEndpoint setDataPort(int dataPort)
-    {
-        this.dataPort = dataPort;
-        return this;
-    }
-
-    // roles
-
-    public Roles getRoles()
-    {
-        return roles;
-    }
-
-    public DrillbitEndpoint setRoles(Roles roles)
-    {
-        this.roles = roles;
-        return this;
-    }
-
-    // version
-
-    public String getVersion()
-    {
-        return version;
-    }
-
-    public DrillbitEndpoint setVersion(String version)
-    {
-        this.version = version;
-        return this;
-    }
-
-    // state
-
-    public State getState()
-    {
-        return state == null ? State.STARTUP : state;
-    }
-
-    public DrillbitEndpoint setState(State state)
-    {
-        this.state = state;
-        return this;
-    }
-
-    // httpPort
-
-    public int getHttpPort()
-    {
-        return httpPort;
-    }
-
-    public DrillbitEndpoint setHttpPort(int httpPort)
-    {
-        this.httpPort = httpPort;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<DrillbitEndpoint> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public DrillbitEndpoint newMessage()
-    {
-        return new DrillbitEndpoint();
-    }
-
-    public Class<DrillbitEndpoint> typeClass()
-    {
-        return DrillbitEndpoint.class;
-    }
-
-    public String messageName()
-    {
-        return DrillbitEndpoint.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return DrillbitEndpoint.class.getName();
-    }
-
-    public boolean isInitialized(DrillbitEndpoint message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, DrillbitEndpoint message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.address = input.readString();
-                    break;
-                case 2:
-                    message.userPort = input.readInt32();
-                    break;
-                case 3:
-                    message.controlPort = input.readInt32();
-                    break;
-                case 4:
-                    message.dataPort = input.readInt32();
-                    break;
-                case 5:
-                    message.roles = input.mergeObject(message.roles, Roles.getSchema());
-                    break;
-
-                case 6:
-                    message.version = input.readString();
-                    break;
-                case 7:
-                    message.state = State.valueOf(input.readEnum());
-                    break;
-                case 8:
-                    message.httpPort = input.readInt32();
-                    break;
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, DrillbitEndpoint message) throws IOException
-    {
-        if(message.address != null)
-            output.writeString(1, message.address, false);
-
-        if(message.userPort != 0)
-            output.writeInt32(2, message.userPort, false);
-
-        if(message.controlPort != 0)
-            output.writeInt32(3, message.controlPort, false);
-
-        if(message.dataPort != 0)
-            output.writeInt32(4, message.dataPort, false);
-
-        if(message.roles != null)
-             output.writeObject(5, message.roles, Roles.getSchema(), false);
-
-
-        if(message.version != null)
-            output.writeString(6, message.version, false);
-
-        if(message.state != null)
-             output.writeEnum(7, message.state.number, false);
-
-        if(message.httpPort != 0)
-            output.writeInt32(8, message.httpPort, false);
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "address";
-            case 2: return "userPort";
-            case 3: return "controlPort";
-            case 4: return "dataPort";
-            case 5: return "roles";
-            case 6: return "version";
-            case 7: return "state";
-            case 8: return "httpPort";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("address", 1);
-        __fieldMap.put("userPort", 2);
-        __fieldMap.put("controlPort", 3);
-        __fieldMap.put("dataPort", 4);
-        __fieldMap.put("roles", 5);
-        __fieldMap.put("version", 6);
-        __fieldMap.put("state", 7);
-        __fieldMap.put("httpPort", 8);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/ExceptionWrapper.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/ExceptionWrapper.java
deleted file mode 100644
index de0ee10..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/ExceptionWrapper.java
+++ /dev/null
@@ -1,243 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-import java.util.ArrayList;
-import java.util.List;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class ExceptionWrapper implements Externalizable, Message<ExceptionWrapper>, Schema<ExceptionWrapper>
-{
-
-    public static Schema<ExceptionWrapper> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static ExceptionWrapper getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final ExceptionWrapper DEFAULT_INSTANCE = new ExceptionWrapper();
-
-    
-    private String exceptionClass;
-    private String message;
-    private List<StackTraceElementWrapper> stackTrace;
-    private ExceptionWrapper cause;
-
-    public ExceptionWrapper()
-    {
-        
-    }
-
-    // getters and setters
-
-    // exceptionClass
-
-    public String getExceptionClass()
-    {
-        return exceptionClass;
-    }
-
-    public ExceptionWrapper setExceptionClass(String exceptionClass)
-    {
-        this.exceptionClass = exceptionClass;
-        return this;
-    }
-
-    // message
-
-    public String getMessage()
-    {
-        return message;
-    }
-
-    public ExceptionWrapper setMessage(String message)
-    {
-        this.message = message;
-        return this;
-    }
-
-    // stackTrace
-
-    public List<StackTraceElementWrapper> getStackTraceList()
-    {
-        return stackTrace;
-    }
-
-    public ExceptionWrapper setStackTraceList(List<StackTraceElementWrapper> stackTrace)
-    {
-        this.stackTrace = stackTrace;
-        return this;
-    }
-
-    // cause
-
-    public ExceptionWrapper getCause()
-    {
-        return cause;
-    }
-
-    public ExceptionWrapper setCause(ExceptionWrapper cause)
-    {
-        this.cause = cause;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<ExceptionWrapper> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public ExceptionWrapper newMessage()
-    {
-        return new ExceptionWrapper();
-    }
-
-    public Class<ExceptionWrapper> typeClass()
-    {
-        return ExceptionWrapper.class;
-    }
-
-    public String messageName()
-    {
-        return ExceptionWrapper.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return ExceptionWrapper.class.getName();
-    }
-
-    public boolean isInitialized(ExceptionWrapper message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, ExceptionWrapper message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.exceptionClass = input.readString();
-                    break;
-                case 2:
-                    message.message = input.readString();
-                    break;
-                case 3:
-                    if(message.stackTrace == null)
-                        message.stackTrace = new ArrayList<StackTraceElementWrapper>();
-                    message.stackTrace.add(input.mergeObject(null, StackTraceElementWrapper.getSchema()));
-                    break;
-
-                case 4:
-                    message.cause = input.mergeObject(message.cause, ExceptionWrapper.getSchema());
-                    break;
-
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, ExceptionWrapper message) throws IOException
-    {
-        if(message.exceptionClass != null)
-            output.writeString(1, message.exceptionClass, false);
-
-        if(message.message != null)
-            output.writeString(2, message.message, false);
-
-        if(message.stackTrace != null)
-        {
-            for(StackTraceElementWrapper stackTrace : message.stackTrace)
-            {
-                if(stackTrace != null)
-                    output.writeObject(3, stackTrace, StackTraceElementWrapper.getSchema(), true);
-            }
-        }
-
-
-        if(message.cause != null)
-             output.writeObject(4, message.cause, ExceptionWrapper.getSchema(), false);
-
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "exceptionClass";
-            case 2: return "message";
-            case 3: return "stackTrace";
-            case 4: return "cause";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("exceptionClass", 1);
-        __fieldMap.put("message", 2);
-        __fieldMap.put("stackTrace", 3);
-        __fieldMap.put("cause", 4);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/FinishedReceiver.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/FinishedReceiver.java
deleted file mode 100644
index 24aa562..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/FinishedReceiver.java
+++ /dev/null
@@ -1,189 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class FinishedReceiver implements Externalizable, Message<FinishedReceiver>, Schema<FinishedReceiver>
-{
-
-    public static Schema<FinishedReceiver> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static FinishedReceiver getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final FinishedReceiver DEFAULT_INSTANCE = new FinishedReceiver();
-
-    
-    private FragmentHandle receiver;
-    private FragmentHandle sender;
-
-    public FinishedReceiver()
-    {
-        
-    }
-
-    // getters and setters
-
-    // receiver
-
-    public FragmentHandle getReceiver()
-    {
-        return receiver;
-    }
-
-    public FinishedReceiver setReceiver(FragmentHandle receiver)
-    {
-        this.receiver = receiver;
-        return this;
-    }
-
-    // sender
-
-    public FragmentHandle getSender()
-    {
-        return sender;
-    }
-
-    public FinishedReceiver setSender(FragmentHandle sender)
-    {
-        this.sender = sender;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<FinishedReceiver> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public FinishedReceiver newMessage()
-    {
-        return new FinishedReceiver();
-    }
-
-    public Class<FinishedReceiver> typeClass()
-    {
-        return FinishedReceiver.class;
-    }
-
-    public String messageName()
-    {
-        return FinishedReceiver.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return FinishedReceiver.class.getName();
-    }
-
-    public boolean isInitialized(FinishedReceiver message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, FinishedReceiver message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.receiver = input.mergeObject(message.receiver, FragmentHandle.getSchema());
-                    break;
-
-                case 2:
-                    message.sender = input.mergeObject(message.sender, FragmentHandle.getSchema());
-                    break;
-
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, FinishedReceiver message) throws IOException
-    {
-        if(message.receiver != null)
-             output.writeObject(1, message.receiver, FragmentHandle.getSchema(), false);
-
-
-        if(message.sender != null)
-             output.writeObject(2, message.sender, FragmentHandle.getSchema(), false);
-
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "receiver";
-            case 2: return "sender";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("receiver", 1);
-        __fieldMap.put("sender", 2);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/FragmentHandle.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/FragmentHandle.java
deleted file mode 100644
index f7088e9..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/FragmentHandle.java
+++ /dev/null
@@ -1,233 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class FragmentHandle implements Externalizable, Message<FragmentHandle>, Schema<FragmentHandle>
-{
-
-    public static Schema<FragmentHandle> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static FragmentHandle getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final FragmentHandle DEFAULT_INSTANCE = new FragmentHandle();
-
-    
-    private QueryId queryId;
-    private int majorFragmentId;
-    private int minorFragmentId;
-    private QueryId parentQueryId;
-
-    public FragmentHandle()
-    {
-        
-    }
-
-    // getters and setters
-
-    // queryId
-
-    public QueryId getQueryId()
-    {
-        return queryId;
-    }
-
-    public FragmentHandle setQueryId(QueryId queryId)
-    {
-        this.queryId = queryId;
-        return this;
-    }
-
-    // majorFragmentId
-
-    public int getMajorFragmentId()
-    {
-        return majorFragmentId;
-    }
-
-    public FragmentHandle setMajorFragmentId(int majorFragmentId)
-    {
-        this.majorFragmentId = majorFragmentId;
-        return this;
-    }
-
-    // minorFragmentId
-
-    public int getMinorFragmentId()
-    {
-        return minorFragmentId;
-    }
-
-    public FragmentHandle setMinorFragmentId(int minorFragmentId)
-    {
-        this.minorFragmentId = minorFragmentId;
-        return this;
-    }
-
-    // parentQueryId
-
-    public QueryId getParentQueryId()
-    {
-        return parentQueryId;
-    }
-
-    public FragmentHandle setParentQueryId(QueryId parentQueryId)
-    {
-        this.parentQueryId = parentQueryId;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<FragmentHandle> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public FragmentHandle newMessage()
-    {
-        return new FragmentHandle();
-    }
-
-    public Class<FragmentHandle> typeClass()
-    {
-        return FragmentHandle.class;
-    }
-
-    public String messageName()
-    {
-        return FragmentHandle.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return FragmentHandle.class.getName();
-    }
-
-    public boolean isInitialized(FragmentHandle message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, FragmentHandle message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.queryId = input.mergeObject(message.queryId, QueryId.getSchema());
-                    break;
-
-                case 2:
-                    message.majorFragmentId = input.readInt32();
-                    break;
-                case 3:
-                    message.minorFragmentId = input.readInt32();
-                    break;
-                case 4:
-                    message.parentQueryId = input.mergeObject(message.parentQueryId, QueryId.getSchema());
-                    break;
-
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, FragmentHandle message) throws IOException
-    {
-        if(message.queryId != null)
-             output.writeObject(1, message.queryId, QueryId.getSchema(), false);
-
-
-        if(message.majorFragmentId != 0)
-            output.writeInt32(2, message.majorFragmentId, false);
-
-        if(message.minorFragmentId != 0)
-            output.writeInt32(3, message.minorFragmentId, false);
-
-        if(message.parentQueryId != null)
-             output.writeObject(4, message.parentQueryId, QueryId.getSchema(), false);
-
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "queryId";
-            case 2: return "majorFragmentId";
-            case 3: return "minorFragmentId";
-            case 4: return "parentQueryId";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("queryId", 1);
-        __fieldMap.put("majorFragmentId", 2);
-        __fieldMap.put("minorFragmentId", 3);
-        __fieldMap.put("parentQueryId", 4);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/FragmentRecordBatch.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/FragmentRecordBatch.java
deleted file mode 100644
index 5bad25c..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/FragmentRecordBatch.java
+++ /dev/null
@@ -1,309 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-import java.util.ArrayList;
-import java.util.List;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class FragmentRecordBatch implements Externalizable, Message<FragmentRecordBatch>, Schema<FragmentRecordBatch>
-{
-
-    public static Schema<FragmentRecordBatch> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static FragmentRecordBatch getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final FragmentRecordBatch DEFAULT_INSTANCE = new FragmentRecordBatch();
-
-    
-    private QueryId queryId;
-    private int receivingMajorFragmentId;
-    private List<Integer> receivingMinorFragmentId;
-    private int sendingMajorFragmentId;
-    private int sendingMinorFragmentId;
-    private RecordBatchDef def;
-    private Boolean isLastBatch;
-
-    public FragmentRecordBatch()
-    {
-        
-    }
-
-    // getters and setters
-
-    // queryId
-
-    public QueryId getQueryId()
-    {
-        return queryId;
-    }
-
-    public FragmentRecordBatch setQueryId(QueryId queryId)
-    {
-        this.queryId = queryId;
-        return this;
-    }
-
-    // receivingMajorFragmentId
-
-    public int getReceivingMajorFragmentId()
-    {
-        return receivingMajorFragmentId;
-    }
-
-    public FragmentRecordBatch setReceivingMajorFragmentId(int receivingMajorFragmentId)
-    {
-        this.receivingMajorFragmentId = receivingMajorFragmentId;
-        return this;
-    }
-
-    // receivingMinorFragmentId
-
-    public List<Integer> getReceivingMinorFragmentIdList()
-    {
-        return receivingMinorFragmentId;
-    }
-
-    public FragmentRecordBatch setReceivingMinorFragmentIdList(List<Integer> receivingMinorFragmentId)
-    {
-        this.receivingMinorFragmentId = receivingMinorFragmentId;
-        return this;
-    }
-
-    // sendingMajorFragmentId
-
-    public int getSendingMajorFragmentId()
-    {
-        return sendingMajorFragmentId;
-    }
-
-    public FragmentRecordBatch setSendingMajorFragmentId(int sendingMajorFragmentId)
-    {
-        this.sendingMajorFragmentId = sendingMajorFragmentId;
-        return this;
-    }
-
-    // sendingMinorFragmentId
-
-    public int getSendingMinorFragmentId()
-    {
-        return sendingMinorFragmentId;
-    }
-
-    public FragmentRecordBatch setSendingMinorFragmentId(int sendingMinorFragmentId)
-    {
-        this.sendingMinorFragmentId = sendingMinorFragmentId;
-        return this;
-    }
-
-    // def
-
-    public RecordBatchDef getDef()
-    {
-        return def;
-    }
-
-    public FragmentRecordBatch setDef(RecordBatchDef def)
-    {
-        this.def = def;
-        return this;
-    }
-
-    // isLastBatch
-
-    public Boolean getIsLastBatch()
-    {
-        return isLastBatch;
-    }
-
-    public FragmentRecordBatch setIsLastBatch(Boolean isLastBatch)
-    {
-        this.isLastBatch = isLastBatch;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<FragmentRecordBatch> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public FragmentRecordBatch newMessage()
-    {
-        return new FragmentRecordBatch();
-    }
-
-    public Class<FragmentRecordBatch> typeClass()
-    {
-        return FragmentRecordBatch.class;
-    }
-
-    public String messageName()
-    {
-        return FragmentRecordBatch.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return FragmentRecordBatch.class.getName();
-    }
-
-    public boolean isInitialized(FragmentRecordBatch message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, FragmentRecordBatch message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.queryId = input.mergeObject(message.queryId, QueryId.getSchema());
-                    break;
-
-                case 2:
-                    message.receivingMajorFragmentId = input.readInt32();
-                    break;
-                case 3:
-                    if(message.receivingMinorFragmentId == null)
-                        message.receivingMinorFragmentId = new ArrayList<Integer>();
-                    message.receivingMinorFragmentId.add(input.readInt32());
-                    break;
-                case 4:
-                    message.sendingMajorFragmentId = input.readInt32();
-                    break;
-                case 5:
-                    message.sendingMinorFragmentId = input.readInt32();
-                    break;
-                case 6:
-                    message.def = input.mergeObject(message.def, RecordBatchDef.getSchema());
-                    break;
-
-                case 7:
-                    message.isLastBatch = input.readBool();
-                    break;
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, FragmentRecordBatch message) throws IOException
-    {
-        if(message.queryId != null)
-             output.writeObject(1, message.queryId, QueryId.getSchema(), false);
-
-
-        if(message.receivingMajorFragmentId != 0)
-            output.writeInt32(2, message.receivingMajorFragmentId, false);
-
-        if(message.receivingMinorFragmentId != null)
-        {
-            for(Integer receivingMinorFragmentId : message.receivingMinorFragmentId)
-            {
-                if(receivingMinorFragmentId != null)
-                    output.writeInt32(3, receivingMinorFragmentId, true);
-            }
-        }
-
-        if(message.sendingMajorFragmentId != 0)
-            output.writeInt32(4, message.sendingMajorFragmentId, false);
-
-        if(message.sendingMinorFragmentId != 0)
-            output.writeInt32(5, message.sendingMinorFragmentId, false);
-
-        if(message.def != null)
-             output.writeObject(6, message.def, RecordBatchDef.getSchema(), false);
-
-
-        if(message.isLastBatch != null)
-            output.writeBool(7, message.isLastBatch, false);
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "queryId";
-            case 2: return "receivingMajorFragmentId";
-            case 3: return "receivingMinorFragmentId";
-            case 4: return "sendingMajorFragmentId";
-            case 5: return "sendingMinorFragmentId";
-            case 6: return "def";
-            case 7: return "isLastBatch";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("queryId", 1);
-        __fieldMap.put("receivingMajorFragmentId", 2);
-        __fieldMap.put("receivingMinorFragmentId", 3);
-        __fieldMap.put("sendingMajorFragmentId", 4);
-        __fieldMap.put("sendingMinorFragmentId", 5);
-        __fieldMap.put("def", 6);
-        __fieldMap.put("isLastBatch", 7);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/FragmentState.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/FragmentState.java
deleted file mode 100644
index 1869318..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/FragmentState.java
+++ /dev/null
@@ -1,59 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-public enum FragmentState implements com.dyuproject.protostuff.EnumLite<FragmentState>
-{
-    SENDING(0),
-    AWAITING_ALLOCATION(1),
-    RUNNING(2),
-    FINISHED(3),
-    CANCELLED(4),
-    FAILED(5),
-    CANCELLATION_REQUESTED(6);
-    
-    public final int number;
-    
-    private FragmentState (int number)
-    {
-        this.number = number;
-    }
-    
-    public int getNumber()
-    {
-        return number;
-    }
-    
-    public static FragmentState valueOf(int number)
-    {
-        switch(number) 
-        {
-            case 0: return SENDING;
-            case 1: return AWAITING_ALLOCATION;
-            case 2: return RUNNING;
-            case 3: return FINISHED;
-            case 4: return CANCELLED;
-            case 5: return FAILED;
-            case 6: return CANCELLATION_REQUESTED;
-            default: return null;
-        }
-    }
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/FragmentStatus.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/FragmentStatus.java
deleted file mode 100644
index de916b4..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/FragmentStatus.java
+++ /dev/null
@@ -1,189 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class FragmentStatus implements Externalizable, Message<FragmentStatus>, Schema<FragmentStatus>
-{
-
-    public static Schema<FragmentStatus> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static FragmentStatus getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final FragmentStatus DEFAULT_INSTANCE = new FragmentStatus();
-
-    
-    private MinorFragmentProfile profile;
-    private FragmentHandle handle;
-
-    public FragmentStatus()
-    {
-        
-    }
-
-    // getters and setters
-
-    // profile
-
-    public MinorFragmentProfile getProfile()
-    {
-        return profile;
-    }
-
-    public FragmentStatus setProfile(MinorFragmentProfile profile)
-    {
-        this.profile = profile;
-        return this;
-    }
-
-    // handle
-
-    public FragmentHandle getHandle()
-    {
-        return handle;
-    }
-
-    public FragmentStatus setHandle(FragmentHandle handle)
-    {
-        this.handle = handle;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<FragmentStatus> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public FragmentStatus newMessage()
-    {
-        return new FragmentStatus();
-    }
-
-    public Class<FragmentStatus> typeClass()
-    {
-        return FragmentStatus.class;
-    }
-
-    public String messageName()
-    {
-        return FragmentStatus.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return FragmentStatus.class.getName();
-    }
-
-    public boolean isInitialized(FragmentStatus message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, FragmentStatus message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.profile = input.mergeObject(message.profile, MinorFragmentProfile.getSchema());
-                    break;
-
-                case 2:
-                    message.handle = input.mergeObject(message.handle, FragmentHandle.getSchema());
-                    break;
-
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, FragmentStatus message) throws IOException
-    {
-        if(message.profile != null)
-             output.writeObject(1, message.profile, MinorFragmentProfile.getSchema(), false);
-
-
-        if(message.handle != null)
-             output.writeObject(2, message.handle, FragmentHandle.getSchema(), false);
-
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "profile";
-            case 2: return "handle";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("profile", 1);
-        __fieldMap.put("handle", 2);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/GetCatalogsReq.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/GetCatalogsReq.java
deleted file mode 100644
index 328ad3c..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/GetCatalogsReq.java
+++ /dev/null
@@ -1,165 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class GetCatalogsReq implements Externalizable, Message<GetCatalogsReq>, Schema<GetCatalogsReq>
-{
-
-    public static Schema<GetCatalogsReq> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static GetCatalogsReq getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final GetCatalogsReq DEFAULT_INSTANCE = new GetCatalogsReq();
-
-    
-    private LikeFilter catalogNameFilter;
-
-    public GetCatalogsReq()
-    {
-        
-    }
-
-    // getters and setters
-
-    // catalogNameFilter
-
-    public LikeFilter getCatalogNameFilter()
-    {
-        return catalogNameFilter;
-    }
-
-    public GetCatalogsReq setCatalogNameFilter(LikeFilter catalogNameFilter)
-    {
-        this.catalogNameFilter = catalogNameFilter;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<GetCatalogsReq> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public GetCatalogsReq newMessage()
-    {
-        return new GetCatalogsReq();
-    }
-
-    public Class<GetCatalogsReq> typeClass()
-    {
-        return GetCatalogsReq.class;
-    }
-
-    public String messageName()
-    {
-        return GetCatalogsReq.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return GetCatalogsReq.class.getName();
-    }
-
-    public boolean isInitialized(GetCatalogsReq message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, GetCatalogsReq message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.catalogNameFilter = input.mergeObject(message.catalogNameFilter, LikeFilter.getSchema());
-                    break;
-
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, GetCatalogsReq message) throws IOException
-    {
-        if(message.catalogNameFilter != null)
-             output.writeObject(1, message.catalogNameFilter, LikeFilter.getSchema(), false);
-
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "catalogNameFilter";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("catalogNameFilter", 1);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/GetCatalogsResp.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/GetCatalogsResp.java
deleted file mode 100644
index 23981f3..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/GetCatalogsResp.java
+++ /dev/null
@@ -1,221 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-import java.util.ArrayList;
-import java.util.List;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class GetCatalogsResp implements Externalizable, Message<GetCatalogsResp>, Schema<GetCatalogsResp>
-{
-
-    public static Schema<GetCatalogsResp> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static GetCatalogsResp getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final GetCatalogsResp DEFAULT_INSTANCE = new GetCatalogsResp();
-
-    
-    private RequestStatus status;
-    private List<CatalogMetadata> catalogs;
-    private DrillPBError error;
-
-    public GetCatalogsResp()
-    {
-        
-    }
-
-    // getters and setters
-
-    // status
-
-    public RequestStatus getStatus()
-    {
-        return status == null ? RequestStatus.UNKNOWN_STATUS : status;
-    }
-
-    public GetCatalogsResp setStatus(RequestStatus status)
-    {
-        this.status = status;
-        return this;
-    }
-
-    // catalogs
-
-    public List<CatalogMetadata> getCatalogsList()
-    {
-        return catalogs;
-    }
-
-    public GetCatalogsResp setCatalogsList(List<CatalogMetadata> catalogs)
-    {
-        this.catalogs = catalogs;
-        return this;
-    }
-
-    // error
-
-    public DrillPBError getError()
-    {
-        return error;
-    }
-
-    public GetCatalogsResp setError(DrillPBError error)
-    {
-        this.error = error;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<GetCatalogsResp> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public GetCatalogsResp newMessage()
-    {
-        return new GetCatalogsResp();
-    }
-
-    public Class<GetCatalogsResp> typeClass()
-    {
-        return GetCatalogsResp.class;
-    }
-
-    public String messageName()
-    {
-        return GetCatalogsResp.class.getSimpleName();
-    }
-
-    public String messageFullName()
-    {
-        return GetCatalogsResp.class.getName();
-    }
-
-    public boolean isInitialized(GetCatalogsResp message)
-    {
-        return true;
-    }
-
-    public void mergeFrom(Input input, GetCatalogsResp message) throws IOException
-    {
-        for(int number = input.readFieldNumber(this);; number = input.readFieldNumber(this))
-        {
-            switch(number)
-            {
-                case 0:
-                    return;
-                case 1:
-                    message.status = RequestStatus.valueOf(input.readEnum());
-                    break;
-                case 2:
-                    if(message.catalogs == null)
-                        message.catalogs = new ArrayList<CatalogMetadata>();
-                    message.catalogs.add(input.mergeObject(null, CatalogMetadata.getSchema()));
-                    break;
-
-                case 3:
-                    message.error = input.mergeObject(message.error, DrillPBError.getSchema());
-                    break;
-
-                default:
-                    input.handleUnknownField(number, this);
-            }   
-        }
-    }
-
-
-    public void writeTo(Output output, GetCatalogsResp message) throws IOException
-    {
-        if(message.status != null)
-             output.writeEnum(1, message.status.number, false);
-
-        if(message.catalogs != null)
-        {
-            for(CatalogMetadata catalogs : message.catalogs)
-            {
-                if(catalogs != null)
-                    output.writeObject(2, catalogs, CatalogMetadata.getSchema(), true);
-            }
-        }
-
-
-        if(message.error != null)
-             output.writeObject(3, message.error, DrillPBError.getSchema(), false);
-
-    }
-
-    public String getFieldName(int number)
-    {
-        switch(number)
-        {
-            case 1: return "status";
-            case 2: return "catalogs";
-            case 3: return "error";
-            default: return null;
-        }
-    }
-
-    public int getFieldNumber(String name)
-    {
-        final Integer number = __fieldMap.get(name);
-        return number == null ? 0 : number.intValue();
-    }
-
-    private static final java.util.HashMap<String,Integer> __fieldMap = new java.util.HashMap<String,Integer>();
-    static
-    {
-        __fieldMap.put("status", 1);
-        __fieldMap.put("catalogs", 2);
-        __fieldMap.put("error", 3);
-    }
-    
-}
diff --git a/protocol/src/main/java/org/apache/drill/exec/proto/beans/GetColumnsReq.java b/protocol/src/main/java/org/apache/drill/exec/proto/beans/GetColumnsReq.java
deleted file mode 100644
index 276d162..0000000
--- a/protocol/src/main/java/org/apache/drill/exec/proto/beans/GetColumnsReq.java
+++ /dev/null
@@ -1,237 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Generated by http://code.google.com/p/protostuff/ ... DO NOT EDIT!
-// Generated from protobuf
-
-package org.apache.drill.exec.proto.beans;
-
-import java.io.Externalizable;
-import java.io.IOException;
-import java.io.ObjectInput;
-import java.io.ObjectOutput;
-
-import com.dyuproject.protostuff.GraphIOUtil;
-import com.dyuproject.protostuff.Input;
-import com.dyuproject.protostuff.Message;
-import com.dyuproject.protostuff.Output;
-import com.dyuproject.protostuff.Schema;
-
-public final class GetColumnsReq implements Externalizable, Message<GetColumnsReq>, Schema<GetColumnsReq>
-{
-
-    public static Schema<GetColumnsReq> getSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    public static GetColumnsReq getDefaultInstance()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    static final GetColumnsReq DEFAULT_INSTANCE = new GetColumnsReq();
-
-    
-    private LikeFilter catalogNameFilter;
-    private LikeFilter schemaNameFilter;
-    private LikeFilter tableNameFilter;
-    private LikeFilter columnNameFilter;
-
-    public GetColumnsReq()
-    {
-        
-    }
-
-    // getters and setters
-
-    // catalogNameFilter
-
-    public LikeFilter getCatalogNameFilter()
-    {
-        return catalogNameFilter;
-    }
-
-    public GetColumnsReq setCatalogNameFilter(LikeFilter catalogNameFilter)
-    {
-        this.catalogNameFilter = catalogNameFilter;
-        return this;
-    }
-
-    // schemaNameFilter
-
-    public LikeFilter getSchemaNameFilter()
-    {
-        return schemaNameFilter;
-    }
-
-    public GetColumnsReq setSchemaNameFilter(LikeFilter schemaNameFilter)
-    {
-        this.schemaNameFilter = schemaNameFilter;
-        return this;
-    }
-
-    // tableNameFilter
-
-    public LikeFilter getTableNameFilter()
-    {
-        return tableNameFilter;
-    }
-
-    public GetColumnsReq setTableNameFilter(LikeFilter tableNameFilter)
-    {
-        this.tableNameFilter = tableNameFilter;
-        return this;
-    }
-
-    // columnNameFilter
-
-    public LikeFilter getColumnNameFilter()
-    {
-        return columnNameFilter;
-    }
-
-    public GetColumnsReq setColumnNameFilter(LikeFilter columnNameFilter)
-    {
-        this.columnNameFilter = columnNameFilter;
-        return this;
-    }
-
-    // java serialization
-
-    public void readExternal(ObjectInput in) throws IOException
-    {
-        GraphIOUtil.mergeDelimitedFrom(in, this, this);
-    }
-
-    public void writeExternal(ObjectOutput out) throws IOException
-    {
-        GraphIOUtil.writeDelimitedTo(out, this, this);
-    }
-
-    // message method
-
-    public Schema<GetColumnsReq> cachedSchema()
-    {
-        return DEFAULT_INSTANCE;
-    }
-
-    // schema methods
-
-    public GetColumnsReq newMessage()
-    {
-        return new GetColumnsReq();
... 14930 lines suppressed ...


[drill] 03/03: DRILL-7292: Remove V1 and V2 text readers

Posted by vo...@apache.org.
This is an automated email from the ASF dual-hosted git repository.

volodymyr pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/drill.git

commit f3f7dbd40f5e899f2aacba35db8f50ffedfa9d3d
Author: Paul Rogers <pa...@yahoo.com>
AuthorDate: Wed Jun 12 17:31:56 2019 -0700

    DRILL-7292: Remove V1 and V2 text readers
    
    Drill 1.16 introduced the "V2" text reader based on the row set
    and provided schema mechanisms. V3 was available by system/session
    option as the functionality was considered experimental.
    
    The functionality has now undergone thorough testing. This commit makes
    the V3 text reader available by default, and removes the code for the
    original "V1" and the "new" (compliant, "V2") text reader.
    
    The system/session options that controlled reader selection are retained
    for backward compatibility, but they no longer do anything.
    
    Specific changes:
    
    * Removed the V2 "compliant" text reader.
    * Moved the "V3" to replace the "compliant" version.
    * Renamed the "complaint" package to "reader."
    * Removed the V1 text reader.
    * Moved the V1 text writer (still used with the V2 and V3 readers)
      into a new "writer" package adjacent to the reader.
    * Removed the CSV tests for the V2 reader, including those that
      demonstrated bugs in V2.
    * V2 did not properly handle the quote escape character. One or two unit
      tests depended on the broken behavior. Fixed them for the correct
      behavior.
    * Behavior of "messy quotes" (those that appear in a non-quoted field)
      was undefined for the text reader. Added a test to clearly demonstrate
      the (somewhat odd) behavior. The behavior itself was not changed.
    
    Reran all unit tests to ensure that they work with the now-default V3
    text reader.
    
    closes #1806
---
 .../java/org/apache/drill/exec/ExecConstants.java  |  19 +-
 .../impl/scan/columns/ColumnsArrayParser.java      |   2 +-
 .../exec/store/easy/text/TextFormatPlugin.java     |  57 +-
 .../text/compliant/CompliantTextRecordReader.java  | 256 -------
 .../easy/text/compliant/FieldVarCharOutput.java    | 234 ------
 .../store/easy/text/compliant/HeaderBuilder.java   | 263 -------
 .../easy/text/compliant/RepeatedVarCharOutput.java | 346 ---------
 .../exec/store/easy/text/compliant/TextInput.java  | 365 ---------
 .../exec/store/easy/text/compliant/TextOutput.java |  87 ---
 .../easy/text/compliant/TextParsingContext.java    | 122 ---
 .../easy/text/compliant/TextParsingSettings.java   | 291 -------
 .../exec/store/easy/text/compliant/TextReader.java | 511 -------------
 .../store/easy/text/compliant/package-info.java    |  27 -
 .../v3/StreamFinishedPseudoException.java          |  29 -
 .../{compliant/v3 => reader}/BaseFieldOutput.java  |   2 +-
 .../v3 => reader}/CompliantTextBatchReader.java    |   6 +-
 .../v3 => reader}/FieldVarCharOutput.java          |   2 +-
 .../{compliant/v3 => reader}/HeaderBuilder.java    |   2 +-
 .../v3 => reader}/RepeatedVarCharOutput.java       |   2 +-
 .../StreamFinishedPseudoException.java             |   9 +-
 .../text/{compliant/v3 => reader}/TextInput.java   |   6 +-
 .../text/{compliant/v3 => reader}/TextOutput.java  |   2 +-
 .../v3 => reader}/TextParsingContext.java          |   2 +-
 .../TextParsingSettings.java}                      |   7 +-
 .../text/{compliant/v3 => reader}/TextReader.java  |  27 +-
 .../{compliant/v3 => reader}/package-info.java     |   2 +-
 .../text/writer/TextRecordWriter.java}             |  14 +-
 .../exec/store/text/DrillTextRecordReader.java     | 244 ------
 .../org/apache/drill/TestSelectWithOption.java     |  13 +-
 .../drill/exec/store/TestImplicitFileColumns.java  |  16 +-
 .../store/easy/text/compliant/BaseCsvTest.java     |  10 -
 .../text/compliant}/TestCsvHeader.java             |   2 +-
 .../easy/text/compliant/TestCsvIgnoreHeaders.java  |  25 +-
 .../text/compliant/TestCsvTableProperties.java     |  67 ++
 .../easy/text/compliant/TestCsvWithHeaders.java    | 846 ++++++---------------
 .../easy/text/compliant/TestCsvWithSchema.java     | 188 +----
 .../easy/text/compliant/TestCsvWithoutHeaders.java | 344 +++------
 .../easy/text/compliant/TestHeaderBuilder.java     |   2 +-
 .../easy/text/compliant/TestPartitionRace.java     | 227 +-----
 .../text/compliant}/TestTextColumn.java            |  15 +-
 .../text/compliant/TestTextReader.java}            |  11 +-
 .../text/compliant}/TextRecordReaderTest.java      |   6 +-
 .../src/test/resources/store/text/data/letters.csv |   9 +-
 43 files changed, 517 insertions(+), 4200 deletions(-)

diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/ExecConstants.java b/exec/java-exec/src/main/java/org/apache/drill/exec/ExecConstants.java
index 93c9902..a7bfd96 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/ExecConstants.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/ExecConstants.java
@@ -713,13 +713,28 @@ public final class ExecConstants {
   public static final OptionValidator ENABLE_VERBOSE_ERRORS = new BooleanValidator(ENABLE_VERBOSE_ERRORS_KEY,
       new OptionDescription("Toggles verbose output of executable error messages"));
 
+  /**
+   * Key used in earlier versions to use the original ("V1") text reader. Since at least Drill 1.8
+   * users have used the ("compliant") ("V2") version. Deprecated in Drill 1.17; the "V3" reader
+   * with schema support is always used. Retained for backward compatibility, but does
+   * nothing.
+   */
+  @Deprecated
   public static final String ENABLE_NEW_TEXT_READER_KEY = "exec.storage.enable_new_text_reader";
+  @Deprecated
   public static final OptionValidator ENABLE_NEW_TEXT_READER = new BooleanValidator(ENABLE_NEW_TEXT_READER_KEY,
-      new OptionDescription("Enables the text reader that complies with the RFC 4180 standard for text/csv files."));
+      new OptionDescription("Deprecated. Drill's text reader complies with the RFC 4180 standard for text/csv files."));
 
+  /**
+   * Flag used in Drill 1.16 to select the row-set based ("V3") or the original
+   * "compliant" ("V2") text reader. In Drill 1.17, the "V3" version is always
+   * used. Retained for backward compatibility, but does nothing.
+   */
+  @Deprecated
   public static final String ENABLE_V3_TEXT_READER_KEY = "exec.storage.enable_v3_text_reader";
+  @Deprecated
   public static final OptionValidator ENABLE_V3_TEXT_READER = new BooleanValidator(ENABLE_V3_TEXT_READER_KEY,
-      new OptionDescription("Enables the row set based version of the text/csv reader."));
+      new OptionDescription("Deprecated. The \"V3\" text reader is always used."));
 
   public static final String MIN_READER_WIDTH_KEY = "exec.storage.min_width";
   public static final OptionValidator MIN_READER_WIDTH = new LongValidator(MIN_READER_WIDTH_KEY,
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/scan/columns/ColumnsArrayParser.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/scan/columns/ColumnsArrayParser.java
index 1737271..24a332d 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/scan/columns/ColumnsArrayParser.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/scan/columns/ColumnsArrayParser.java
@@ -24,7 +24,7 @@ import org.apache.drill.exec.physical.impl.scan.project.ScanLevelProjection;
 import org.apache.drill.exec.physical.impl.scan.project.ScanLevelProjection.ScanProjectionParser;
 import org.apache.drill.exec.physical.rowSet.project.RequestedColumnImpl;
 import org.apache.drill.exec.physical.rowSet.project.RequestedTuple.RequestedColumn;
-import org.apache.drill.exec.store.easy.text.compliant.v3.TextReader;
+import org.apache.drill.exec.store.easy.text.reader.TextReader;
 
 /**
  * Parses the `columns` array. Doing so is surprisingly complex.
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/TextFormatPlugin.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/TextFormatPlugin.java
index f12fffb..89ef8f0 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/TextFormatPlugin.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/TextFormatPlugin.java
@@ -47,27 +47,19 @@ import org.apache.drill.exec.proto.UserBitShared.CoreOperatorType;
 import org.apache.drill.exec.record.metadata.TupleMetadata;
 import org.apache.drill.exec.server.DrillbitContext;
 import org.apache.drill.exec.server.options.OptionManager;
-import org.apache.drill.exec.store.RecordReader;
 import org.apache.drill.exec.store.RecordWriter;
-import org.apache.drill.exec.store.dfs.DrillFileSystem;
 import org.apache.drill.exec.store.dfs.FileSelection;
 import org.apache.drill.exec.store.dfs.easy.EasyFormatPlugin;
 import org.apache.drill.exec.store.dfs.easy.EasyGroupScan;
 import org.apache.drill.exec.store.dfs.easy.EasySubScan;
 import org.apache.drill.exec.store.dfs.easy.EasyWriter;
-import org.apache.drill.exec.store.dfs.easy.FileWork;
-import org.apache.drill.exec.store.easy.text.compliant.CompliantTextRecordReader;
-import org.apache.drill.exec.store.easy.text.compliant.TextParsingSettings;
-import org.apache.drill.exec.store.easy.text.compliant.v3.CompliantTextBatchReader;
-import org.apache.drill.exec.store.easy.text.compliant.v3.TextParsingSettingsV3;
+import org.apache.drill.exec.store.easy.text.reader.CompliantTextBatchReader;
+import org.apache.drill.exec.store.easy.text.reader.TextParsingSettings;
+import org.apache.drill.exec.store.easy.text.writer.TextRecordWriter;
 import org.apache.drill.exec.store.schedule.CompleteFileWork;
-import org.apache.drill.exec.store.text.DrillTextRecordReader;
-import org.apache.drill.exec.store.text.DrillTextRecordWriter;
 import org.apache.drill.exec.vector.accessor.convert.AbstractConvertFromString;
 import org.apache.drill.shaded.guava.com.google.common.collect.ImmutableList;
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.mapred.FileSplit;
 
 import com.fasterxml.jackson.annotation.JsonIgnore;
 import com.fasterxml.jackson.annotation.JsonInclude;
@@ -212,9 +204,9 @@ public class TextFormatPlugin extends EasyFormatPlugin<TextFormatPlugin.TextForm
    */
   private static class ColumnsReaderFactory extends FileReaderFactory {
 
-    private final TextParsingSettingsV3 settings;
+    private final TextParsingSettings settings;
 
-    public ColumnsReaderFactory(TextParsingSettingsV3 settings) {
+    public ColumnsReaderFactory(TextParsingSettings settings) {
       this.settings = settings;
     }
 
@@ -246,9 +238,7 @@ public class TextFormatPlugin extends EasyFormatPlugin<TextFormatPlugin.TextForm
     config.defaultName = PLUGIN_NAME;
     config.readerOperatorType = CoreOperatorType.TEXT_SUB_SCAN_VALUE;
     config.writerOperatorType = CoreOperatorType.TEXT_WRITER_VALUE;
-
-    // Uncomment this, and remove useEnhancedScan(), when v2 is retired
-    //config.useEnhancedScan = true;
+    config.useEnhancedScan = true;
     return config;
   }
 
@@ -271,40 +261,11 @@ public class TextFormatPlugin extends EasyFormatPlugin<TextFormatPlugin.TextForm
   }
 
   @Override
-  protected boolean useEnhancedScan(OptionManager options) {
-    // Create the "legacy", "V2" reader or the new "V3" version based on
-    // the result set loader. This code should be temporary: the two
-    // readers provide identical functionality for the user; only the
-    // internals differ.
-    return options.getBoolean(ExecConstants.ENABLE_V3_TEXT_READER_KEY);
-  }
-
-  // TODO: Remove this once the V2 reader is removed.
-  @Override
-  public RecordReader getRecordReader(FragmentContext context,
-                                      DrillFileSystem dfs,
-                                      FileWork fileWork,
-                                      List<SchemaPath> columns,
-                                      String userName) {
-    Path path = dfs.makeQualified(fileWork.getPath());
-    FileSplit split = new FileSplit(path, fileWork.getStart(), fileWork.getLength(), new String[]{""});
-
-    if (context.getOptions().getBoolean(ExecConstants.ENABLE_NEW_TEXT_READER_KEY)) {
-      TextParsingSettings settings = new TextParsingSettings();
-      settings.set(formatConfig);
-      return new CompliantTextRecordReader(split, dfs, settings, columns);
-    } else {
-      char delim = formatConfig.getFieldDelimiter();
-      return new DrillTextRecordReader(split, dfs.getConf(), context, delim, columns);
-    }
-  }
-
-  @Override
   protected FileScanBuilder frameworkBuilder(
       OptionManager options, EasySubScan scan) throws ExecutionSetupException {
     ColumnsScanBuilder builder = new ColumnsScanBuilder();
-    TextParsingSettingsV3 settings =
-        new TextParsingSettingsV3(getConfig(), scan, options);
+    TextParsingSettings settings =
+        new TextParsingSettings(getConfig(), scan, options);
     builder.setReaderFactory(new ColumnsReaderFactory(settings));
 
     // If this format has no headers, or wants to skip them,
@@ -359,7 +320,7 @@ public class TextFormatPlugin extends EasyFormatPlugin<TextFormatPlugin.TextForm
     options.put("separator", getConfig().getFieldDelimiterAsString());
     options.put("extension", getConfig().getExtensions().get(0));
 
-    RecordWriter recordWriter = new DrillTextRecordWriter(
+    RecordWriter recordWriter = new TextRecordWriter(
         context.getAllocator(), writer.getStorageStrategy(), writer.getFormatPlugin().getFsConf());
     recordWriter.init(options);
 
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/CompliantTextRecordReader.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/CompliantTextRecordReader.java
deleted file mode 100644
index 0b091f8..0000000
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/CompliantTextRecordReader.java
+++ /dev/null
@@ -1,256 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.drill.exec.store.easy.text.compliant;
-
-import java.io.IOException;
-import java.io.InputStream;
-import java.util.List;
-
-import javax.annotation.Nullable;
-
-import org.apache.drill.common.exceptions.ExecutionSetupException;
-import org.apache.drill.common.exceptions.UserException;
-import org.apache.drill.common.expression.SchemaPath;
-import org.apache.drill.exec.exception.SchemaChangeException;
-import org.apache.drill.exec.ops.OperatorContext;
-import org.apache.drill.exec.physical.impl.OutputMutator;
-import org.apache.drill.exec.store.AbstractRecordReader;
-import org.apache.drill.exec.store.dfs.DrillFileSystem;
-import org.apache.hadoop.mapred.FileSplit;
-
-import org.apache.drill.shaded.guava.com.google.common.base.Predicate;
-import org.apache.drill.shaded.guava.com.google.common.collect.ImmutableList;
-import org.apache.drill.shaded.guava.com.google.common.collect.Iterables;
-import com.univocity.parsers.common.TextParsingException;
-
-import io.netty.buffer.DrillBuf;
-
-// New text reader, complies with the RFC 4180 standard for text/csv files
-public class CompliantTextRecordReader extends AbstractRecordReader {
-  static final org.slf4j.Logger logger = org.slf4j.LoggerFactory.getLogger(CompliantTextRecordReader.class);
-
-  private static final int MAX_RECORDS_PER_BATCH = 8096;
-  private static final int READ_BUFFER = 1024*1024;
-  private static final int WHITE_SPACE_BUFFER = 64*1024;
-  // When no named column is required, ask SCAN to return a DEFAULT column.
-  // If such column does not exist, it will be returned as a nullable-int column.
-  private static final List<SchemaPath> DEFAULT_NAMED_TEXT_COLS_TO_READ =
-      ImmutableList.of(SchemaPath.getSimplePath("_DEFAULT_COL_TO_READ_"));
-
-  // settings to be used while parsing
-  private TextParsingSettings settings;
-  // Chunk of the file to be read by this reader
-  private FileSplit split;
-  // text reader implementation
-  private TextReader reader;
-  // input buffer
-  private DrillBuf readBuffer;
-  // working buffer to handle whitespace
-  private DrillBuf whitespaceBuffer;
-  private DrillFileSystem dfs;
-  // operator context for OutputMutator
-  private OperatorContext oContext;
-
-  public CompliantTextRecordReader(FileSplit split, DrillFileSystem dfs, TextParsingSettings settings, List<SchemaPath> columns) {
-    this.split = split;
-    this.settings = settings;
-    this.dfs = dfs;
-    setColumns(columns);
-  }
-
-  // checks to see if we are querying all columns(star) or individual columns
-  @Override
-  public boolean isStarQuery() {
-    if(settings.isUseRepeatedVarChar()) {
-      return super.isStarQuery() || Iterables.tryFind(getColumns(), new Predicate<SchemaPath>() {
-        @Override
-        public boolean apply(@Nullable SchemaPath path) {
-          return path.equals(RepeatedVarCharOutput.COLUMNS);
-        }
-      }).isPresent();
-    }
-    return super.isStarQuery();
-  }
-
-  /**
-   * Returns list of default columns to read to replace empty list of columns.
-   * For text files without headers returns "columns[0]".
-   * Text files with headers do not support columns syntax,
-   * so when header extraction is enabled, returns fake named column "_DEFAULT_COL_TO_READ_".
-   *
-   * @return list of default columns to read
-   */
-  @Override
-  protected List<SchemaPath> getDefaultColumnsToRead() {
-    if (settings.isHeaderExtractionEnabled()) {
-      return DEFAULT_NAMED_TEXT_COLS_TO_READ;
-    }
-    return DEFAULT_TEXT_COLS_TO_READ;
-  }
-
-  /**
-   * Performs the initial setup required for the record reader.
-   * Initializes the input stream, handling of the output record batch
-   * and the actual reader to be used.
-   * @param context  operator context from which buffer's will be allocated and managed
-   * @param outputMutator  Used to create the schema in the output record batch
-   * @throws ExecutionSetupException
-   */
-  @Override
-  public void setup(OperatorContext context, OutputMutator outputMutator) throws ExecutionSetupException {
-
-    oContext = context;
-    // Note: DO NOT use managed buffers here. They remain in existence
-    // until the fragment is shut down. The buffers here are large.
-    // If we scan 1000 files, and allocate 1 MB for each, we end up
-    // holding onto 1 GB of memory in managed buffers.
-    // Instead, we allocate the buffers explicitly, and must free
-    // them.
-//    readBuffer = context.getManagedBuffer(READ_BUFFER);
-//    whitespaceBuffer = context.getManagedBuffer(WHITE_SPACE_BUFFER);
-    readBuffer = context.getAllocator().buffer(READ_BUFFER);
-    whitespaceBuffer = context.getAllocator().buffer(WHITE_SPACE_BUFFER);
-
-    // setup Output, Input, and Reader
-    try {
-      TextOutput output = null;
-      TextInput input = null;
-      InputStream stream = null;
-
-      // setup Output using OutputMutator
-      if (settings.isHeaderExtractionEnabled()){
-        //extract header and use that to setup a set of VarCharVectors
-        String [] fieldNames = extractHeader();
-        output = new FieldVarCharOutput(outputMutator, fieldNames, getColumns(), isStarQuery());
-      } else {
-        //simply use RepeatedVarCharVector
-        output = new RepeatedVarCharOutput(outputMutator, getColumns(), isStarQuery());
-      }
-
-      // setup Input using InputStream
-      logger.trace("Opening file {}", split.getPath());
-      stream = dfs.openPossiblyCompressedStream(split.getPath());
-      input = new TextInput(settings, stream, readBuffer, split.getStart(), split.getStart() + split.getLength());
-
-      // setup Reader using Input and Output
-      reader = new TextReader(settings, input, output, whitespaceBuffer);
-      reader.start();
-
-    } catch (SchemaChangeException | IOException e) {
-      throw new ExecutionSetupException(String.format("Failure while setting up text reader for file %s", split.getPath()), e);
-    } catch (IllegalArgumentException e) {
-      throw UserException.dataReadError(e).addContext("File Path", split.getPath().toString()).build(logger);
-    }
-  }
-
-  /**
-   * This method is responsible to implement logic for extracting header from text file
-   * Currently it is assumed to be first line if headerExtractionEnabled is set to true
-   * TODO: enhance to support more common header patterns
-   * @return field name strings
-   */
-  private String [] extractHeader() throws SchemaChangeException, IOException, ExecutionSetupException{
-    assert (settings.isHeaderExtractionEnabled());
-    assert (oContext != null);
-
-    // don't skip header in case skipFirstLine is set true
-    settings.setSkipFirstLine(false);
-
-    HeaderBuilder hOutput = new HeaderBuilder();
-
-    // setup Input using InputStream
-    // we should read file header irrespective of split given given to this reader
-    InputStream hStream = dfs.openPossiblyCompressedStream(split.getPath());
-    TextInput hInput = new TextInput(settings,  hStream, oContext.getManagedBuffer(READ_BUFFER), 0, split.getLength());
-
-    // setup Reader using Input and Output
-    this.reader = new TextReader(settings, hInput, hOutput, oContext.getManagedBuffer(WHITE_SPACE_BUFFER));
-    reader.start();
-
-    // extract first row only
-    reader.parseNext();
-
-    // grab the field names from output
-    String [] fieldNames = hOutput.getHeaders();
-
-    // cleanup and set to skip the first line next time we read input
-    reader.close();
-    settings.setSkipFirstLine(true);
-
-    return fieldNames;
-  }
-
-  /**
-   * Generates the next record batch
-   * @return  number of records in the batch
-   *
-   */
-  @Override
-  public int next() {
-    reader.resetForNextBatch();
-    int cnt = 0;
-
-    try{
-      while(cnt < MAX_RECORDS_PER_BATCH && reader.parseNext()){
-        cnt++;
-      }
-      reader.finishBatch();
-      return cnt;
-    } catch (IOException | TextParsingException e) {
-      throw UserException.dataReadError(e)
-          .addContext("Failure while reading file %s. Happened at or shortly before byte position %d.",
-              split.getPath(), reader.getPos())
-          .build(logger);
-    }
-  }
-
-  /**
-   * Cleanup state once we are finished processing all the records.
-   * This would internally close the input stream we are reading from.
-   */
-  @Override
-  public void close() {
-
-    // Release the buffers allocated above. Double-check to handle
-    // unexpected multiple calls to close().
-
-    if (readBuffer != null) {
-      readBuffer.release();
-      readBuffer = null;
-    }
-    if (whitespaceBuffer != null) {
-      whitespaceBuffer.release();
-      whitespaceBuffer = null;
-    }
-    try {
-      if (reader != null) {
-        reader.close();
-        reader = null;
-      }
-    } catch (IOException e) {
-      logger.warn("Exception while closing stream.", e);
-    }
-  }
-
-  @Override
-  public String toString() {
-    return "CompliantTextRecordReader[File=" + split.getPath()
-        + ", reader=" + reader
-        + "]";
-  }
-}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/FieldVarCharOutput.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/FieldVarCharOutput.java
deleted file mode 100644
index 9734c88..0000000
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/FieldVarCharOutput.java
+++ /dev/null
@@ -1,234 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.drill.exec.store.easy.text.compliant;
-
-import java.util.ArrayList;
-import java.util.Arrays;
-import java.util.Collection;
-import java.util.Collections;
-import java.util.List;
-import java.util.Map;
-
-import org.apache.drill.common.exceptions.UserException;
-import org.apache.drill.common.expression.SchemaPath;
-import org.apache.drill.common.map.CaseInsensitiveMap;
-import org.apache.drill.common.types.TypeProtos;
-import org.apache.drill.common.types.Types;
-import org.apache.drill.exec.exception.SchemaChangeException;
-import org.apache.drill.exec.physical.impl.OutputMutator;
-import org.apache.drill.exec.record.MaterializedField;
-import org.apache.drill.exec.vector.VarCharVector;
-
-/**
- * Class is responsible for generating record batches for text file inputs. We generate
- * a record batch with a set of varchar vectors. A varchar vector contains all the field
- * values for a given column. Each record is a single value within each vector of the set.
- */
-class FieldVarCharOutput extends TextOutput {
-
-  static final org.slf4j.Logger logger = org.slf4j.LoggerFactory.getLogger(FieldVarCharOutput.class);
-  static final String COL_NAME = "columns";
-
-  // array of output vector
-  private final VarCharVector [] vectors;
-  // boolean array indicating which fields are selected (if star query entire array is set to true)
-  private final boolean[] selectedFields;
-  // current vector to which field will be added
-  private VarCharVector currentVector;
-  // track which field is getting appended
-  private int currentFieldIndex = -1;
-  // track chars within field
-  private int currentDataPointer = 0;
-  // track if field is still getting appended
-  private boolean fieldOpen = true;
-  // holds chars for a field
-  private byte[] fieldBytes;
-
-  private boolean collect = true;
-  private boolean rowHasData= false;
-  private static final int MAX_FIELD_LENGTH = 1024 * 64;
-  private int recordCount = 0;
-  private int maxField = 0;
-  private int[] nullCols;
-  private byte nullValue[] = new byte[0];
-
-  /**
-   * We initialize and add the varchar vector for each incoming field in this
-   * constructor.
-   * @param outputMutator  Used to create/modify schema
-   * @param fieldNames Incoming field names
-   * @param columns  List of columns selected in the query
-   * @param isStarQuery  boolean to indicate if all fields are selected or not
-   * @throws SchemaChangeException
-   */
-  public FieldVarCharOutput(OutputMutator outputMutator, String [] fieldNames, Collection<SchemaPath> columns, boolean isStarQuery) throws SchemaChangeException {
-
-    int totalFields = fieldNames.length;
-    List<String> outputColumns = new ArrayList<>(Arrays.asList(fieldNames));
-    List<Integer> nullColumns = new ArrayList<>();
-
-    if (isStarQuery) {
-      maxField = totalFields - 1;
-      this.selectedFields = new boolean[totalFields];
-      Arrays.fill(selectedFields, true);
-    } else {
-      List<Integer> columnIds = new ArrayList<Integer>();
-      Map<String, Integer> headers = CaseInsensitiveMap.newHashMap();
-      for (int i = 0; i < fieldNames.length; i++) {
-        headers.put(fieldNames[i], i);
-      }
-
-      for (SchemaPath path : columns) {
-        int index;
-        String pathStr = path.getRootSegment().getPath();
-        if (pathStr.equals(COL_NAME) && path.getRootSegment().getChild() != null) {
-          //TODO: support both field names and columns index along with predicate pushdown
-          throw UserException
-              .unsupportedError()
-              .message("With extractHeader enabled, only header names are supported")
-              .addContext("column name", pathStr)
-              .addContext("column index", path.getRootSegment().getChild())
-              .build(logger);
-        } else {
-          Integer value = headers.get(pathStr);
-          if (value == null) {
-            // found col that is not a part of fieldNames, add it
-            // this col might be part of some another scanner
-            index = totalFields++;
-            outputColumns.add(pathStr);
-            nullColumns.add(index);
-          } else {
-            index = value;
-          }
-        }
-        columnIds.add(index);
-      }
-      Collections.sort(columnIds);
-
-      this.selectedFields = new boolean[totalFields];
-      for(Integer i : columnIds) {
-        selectedFields[i] = true;
-        maxField = i;
-      }
-    }
-
-    this.vectors = new VarCharVector[totalFields];
-
-    for (int i = 0; i <= maxField; i++) {
-      if (selectedFields[i]) {
-        MaterializedField field = MaterializedField.create(outputColumns.get(i), Types.required(TypeProtos.MinorType.VARCHAR));
-        this.vectors[i] = outputMutator.addField(field, VarCharVector.class);
-      }
-    }
-
-    this.fieldBytes = new byte[MAX_FIELD_LENGTH];
-
-    // Keep track of the null columns to be filled in.
-
-    nullCols = new int[nullColumns.size()];
-    for (int i = 0; i < nullCols.length; i++) {
-      nullCols[i] = nullColumns.get(i);
-    }
-  }
-
-  /**
-   * Start a new record batch. Resets all pointers
-   */
-  @Override
-  public void startBatch() {
-    recordCount = 0;
-    currentFieldIndex= -1;
-    collect = true;
-    fieldOpen = false;
-  }
-
-  @Override
-  public void startField(int index) {
-    currentFieldIndex = index;
-    currentDataPointer = 0;
-    fieldOpen = true;
-    collect = selectedFields[index];
-    currentVector = vectors[index];
-  }
-
-  @Override
-  public void append(byte data) {
-    if (!collect) {
-      return;
-    }
-
-    if (currentDataPointer >= MAX_FIELD_LENGTH -1) {
-      throw UserException
-          .unsupportedError()
-          .message("Trying to write something big in a column")
-          .addContext("columnIndex", currentFieldIndex)
-          .addContext("Limit", MAX_FIELD_LENGTH)
-          .build(logger);
-    }
-
-    fieldBytes[currentDataPointer++] = data;
-  }
-
-  @Override
-  public boolean endField() {
-    fieldOpen = false;
-
-    if (collect) {
-      assert currentVector != null;
-      currentVector.getMutator().setSafe(recordCount, fieldBytes, 0, currentDataPointer);
-    }
-
-    if (currentDataPointer > 0) {
-      this.rowHasData = true;
-    }
-
-    return currentFieldIndex < maxField;
-  }
-
-  @Override
-  public boolean endEmptyField() {
-    return endField();
-  }
-
- @Override
-  public void finishRecord() {
-    if (fieldOpen){
-      endField();
-    }
-
-    // Fill in null (really empty) values.
-
-    for (int i = 0; i < nullCols.length; i++) {
-      vectors[nullCols[i]].getMutator().setSafe(recordCount, nullValue, 0, 0);
-    }
-    recordCount++;
-  }
-
-  @Override
-  public void finishBatch() { }
-
-  @Override
-  public long getRecordCount() {
-    return recordCount;
-  }
-
-  @Override
-  public boolean rowHasData() {
-    return this.rowHasData;
-  }
-}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/HeaderBuilder.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/HeaderBuilder.java
deleted file mode 100644
index 6dfc528..0000000
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/HeaderBuilder.java
+++ /dev/null
@@ -1,263 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.drill.exec.store.easy.text.compliant;
-
-import java.nio.BufferOverflowException;
-import java.nio.ByteBuffer;
-import java.util.ArrayList;
-import java.util.HashSet;
-import java.util.List;
-import java.util.Set;
-
-import org.apache.drill.common.exceptions.UserException;
-
-import org.apache.drill.shaded.guava.com.google.common.base.Charsets;
-
-/**
- * Text output that implements a header reader/parser.
- * The caller parses out the characters of each header;
- * this class assembles UTF-8 bytes into Unicode characters,
- * fixes invalid characters (those not legal for SQL symbols),
- * and maps duplicate names to unique names.
- * <p>
- * That is, this class is as permissive as possible with file
- * headers to avoid spurious query failures for trivial reasons.
- */
-
-// Note: this class uses Java heap strings and the usual Java
-// convenience classes. Since we do heavy Unicode string operations,
-// and read a single row, there is no good reason to try to use
-// value vectors and direct memory for this task.
-
-public class HeaderBuilder extends TextOutput {
-
-  /**
-   * Maximum Drill symbol length, as enforced for headers.
-   * @see <a href="https://drill.apache.org/docs/lexical-structure/#identifier">
-   * identifier documentation</a>
-   */
-  // TODO: Replace with the proper constant, if available
-  public static final int MAX_HEADER_LEN = 1024;
-
-  /**
-   * Prefix used to replace non-alphabetic characters at the start of
-   * a column name. For example, $foo becomes col_foo. Used
-   * because SQL does not allow _foo.
-   */
-
-  public static final String COLUMN_PREFIX = "col_";
-
-  /**
-   * Prefix used to create numbered columns for missing
-   * headers. Typical names: column_1, column_2, ...
-   */
-
-  public static final String ANONYMOUS_COLUMN_PREFIX = "column_";
-
-  public final List<String> headers = new ArrayList<>();
-  public final ByteBuffer currentField = ByteBuffer.allocate(MAX_HEADER_LEN);
-
-  @Override
-  public void startField(int index) {
-    currentField.clear();
-  }
-
-  @Override
-  public boolean endField() {
-    String header = new String(currentField.array(), 0, currentField.position(), Charsets.UTF_8);
-    header = validateSymbol(header);
-    headers.add(header);
-    return true;
-  }
-
-  @Override
-  public boolean endEmptyField() {
-
-    // Empty header will be rewritten to "column_<n>".
-
-    return endField();
-  }
-
-  /**
-   * Validate the header name according to the SQL lexical rules.
-   * @see <a href="https://drill.apache.org/docs/lexical-structure/#identifier">
-   * identifier documentation</a>
-   * @param header the header name to validate
-   */
-
-  // TODO: Replace with existing code, if any.
-  private String validateSymbol(String header) {
-    header = header.trim();
-
-    // To avoid unnecessary query failures, just make up a column name
-    // if the name is missing or all blanks.
-
-    if (header.isEmpty()) {
-      return ANONYMOUS_COLUMN_PREFIX + (headers.size() + 1);
-    }
-    if (! Character.isAlphabetic(header.charAt(0))) {
-      return rewriteHeader(header);
-    }
-    for (int i = 1; i < header.length(); i++) {
-      char ch = header.charAt(i);
-      if (! Character.isAlphabetic(ch)  &&
-          ! Character.isDigit(ch)  &&  ch != '_') {
-        return rewriteHeader(header);
-      }
-    }
-    return header;
-  }
-
-  /**
-   * Given an invalid header, rewrite it to replace illegal characters
-   * with valid ones. The header won't be what the user specified,
-   * but it will be a valid SQL identifier. This solution avoids failing
-   * queries due to corrupted or invalid header data.
-   * <p>
-   * Names with invalid first characters are mapped to "col_". Example:
-   * $foo maps to col_foo. If the only character is non-alphabetic, treat
-   * the column as anonymous and create a generic name: column_4, etc.
-   * <p>
-   * This mapping could create a column that exceeds the maximum length
-   * of 1024. Since that is not really a hard limit, we just live with the
-   * extra few characters.
-   *
-   * @param header the original header
-   * @return the rewritten header, valid for SQL
-   */
-
-  private String rewriteHeader(String header) {
-    final StringBuilder buf = new StringBuilder();
-
-    // If starts with non-alphabetic, can't map the character to
-    // underscore, so just tack on a prefix.
-
-    char ch = header.charAt(0);
-    if (Character.isAlphabetic(ch)) {
-      buf.append(ch);
-    } else if (Character.isDigit(ch)) {
-      buf.append(COLUMN_PREFIX);
-      buf.append(ch);
-
-      // For the strange case of only one character, format
-      // the same as an empty header.
-
-    } else if (header.length() == 1) {
-      return ANONYMOUS_COLUMN_PREFIX + (headers.size() + 1);
-    } else {
-      buf.append(COLUMN_PREFIX);
-    }
-
-    // Convert all remaining invalid characters to underscores
-
-    for (int i = 1; i < header.length(); i++) {
-      ch = header.charAt(i);
-      if (Character.isAlphabetic(ch)  ||
-          Character.isDigit(ch)  ||  ch == '_') {
-        buf.append(ch);
-      } else {
-        buf.append("_");
-      }
-    }
-    return buf.toString();
-  }
-
-  @Override
-  public void append(byte data) {
-
-    // Ensure the data fits. Note that, if the name is Unicode, the actual
-    // number of characters might be less than the limit even though the
-    // byte count exceeds the limit. Fixing this, in general, would require
-    // a buffer four times larger, so we leave that as a later improvement
-    // if ever needed.
-
-    try {
-      currentField.put(data);
-    } catch (BufferOverflowException e) {
-      throw UserException.dataReadError()
-        .message("Column exceeds maximum length of %d", MAX_HEADER_LEN)
-        .build(logger);
-    }
-  }
-
-  @Override
-  public void finishRecord() {
-    if (headers.isEmpty()) {
-      throw UserException.dataReadError()
-        .message("The file must define at least one header.")
-        .build(logger);
-    }
-
-    // Force headers to be unique.
-
-    final Set<String> idents = new HashSet<String>();
-    for (int i = 0; i < headers.size(); i++) {
-      String header = headers.get(i);
-      String key = header.toLowerCase();
-
-      // Is the header a duplicate?
-
-      if (idents.contains(key)) {
-
-        // Make header unique by appending a suffix.
-        // This loop must end because we have a finite
-        // number of headers.
-        // The original column is assumed to be "1", so
-        // the first duplicate is "2", and so on.
-        // Note that this will map columns of the form:
-        // "col,col,col_2,col_2_2" to
-        // "col", "col_2", "col_2_2", "col_2_2_2".
-        // No mapping scheme is perfect...
-
-        for (int l = 2;; l++) {
-          final String rewritten = header + "_" + l;
-          key = rewritten.toLowerCase();
-          if (! idents.contains(key)) {
-            headers.set(i, rewritten);
-            break;
-          }
-        }
-      }
-      idents.add(key);
-    }
-  }
-
-  @Override
-  public long getRecordCount() { return 1; }
-
-  @Override
-  public void startBatch() { }
-
-  @Override
-  public void finishBatch() { }
-
-  @Override
-  public boolean rowHasData() {
-    return ! headers.isEmpty();
-  }
-
-  public String[] getHeaders() {
-
-    // Just return the headers: any needed checks were done in
-    // finishRecord()
-
-    final String array[] = new String[headers.size()];
-    return headers.toArray(array);
-  }
-
-}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/RepeatedVarCharOutput.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/RepeatedVarCharOutput.java
deleted file mode 100644
index c8a681e..0000000
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/RepeatedVarCharOutput.java
+++ /dev/null
@@ -1,346 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.drill.exec.store.easy.text.compliant;
-
-import io.netty.buffer.DrillBuf;
-import io.netty.util.internal.PlatformDependent;
-
-import java.util.ArrayList;
-import java.util.Arrays;
-import java.util.Collection;
-import java.util.Collections;
-import java.util.List;
-
-import org.apache.drill.common.exceptions.ExecutionSetupException;
-import org.apache.drill.common.expression.SchemaPath;
-import org.apache.drill.common.types.TypeProtos;
-import org.apache.drill.common.types.Types;
-import org.apache.drill.exec.exception.SchemaChangeException;
-import org.apache.drill.exec.physical.impl.OutputMutator;
-import org.apache.drill.exec.record.MaterializedField;
-import org.apache.drill.exec.vector.RepeatedVarCharVector;
-
-import org.apache.drill.shaded.guava.com.google.common.base.Preconditions;
-
-/**
- * Class is responsible for generating record batches for text file inputs. We generate
- * a record batch with a single vector of type repeated varchar vector. Each record is a single
- * value within the vector containing all the fields in the record as individual array elements.
- */
-public class RepeatedVarCharOutput extends TextOutput {
-  static final org.slf4j.Logger logger = org.slf4j.LoggerFactory.getLogger(RepeatedVarCharOutput.class);
-
-  static final String COL_NAME = "columns";
-  static final SchemaPath COLUMNS = SchemaPath.getSimplePath("columns");
-  public static final int MAXIMUM_NUMBER_COLUMNS = 64 * 1024;
-
-  // output vector
-  private final RepeatedVarCharVector vector;
-
-  // mutator for the output vector
-  private final RepeatedVarCharVector.Mutator mutator;
-
-  // boolean array indicating which fields are selected (if star query entire array is set to true)
-  private final boolean[] collectedFields;
-
-  // pointer to keep track of the offsets per record
-  private long repeatedOffset;
-
-  // pointer to keep track of the original offsets per record
-  private long repeatedOffsetOriginal;
-
-  // pointer to end of the offset buffer
-  private long repeatedOffsetMax;
-
-  // pointer to the start of the actual data buffer
-  private long characterDataOriginal;
-
-  // pointer to the current location of the data buffer
-  private long characterData;
-
-  // pointer to the end of the data buffer
-  private long characterDataMax;
-
-  // current pointer into the buffer that keeps track of the length of individual fields
-  private long charLengthOffset;
-
-  // pointer to the start of the length buffer
-  private long charLengthOffsetOriginal;
-
-  // pointer to the end of length buffer
-  private long charLengthOffsetMax;
-
-  // pointer to the beginning of the record
-  private long recordStart;
-
-  // total number of records processed (across batches)
-  private long recordCount;
-
-  // number of records processed in this current batch
-  private int batchIndex;
-
-  // current index of the field being processed within the record
-  private int fieldIndex = -1;
-
-  /* boolean to indicate if we are currently appending data to the output vector
-   * Its set to false when we have hit out of memory or we are not interested in
-   * the particular field
-   */
-  private boolean collect;
-
-  // are we currently appending to a field
-  private boolean fieldOpen;
-
-  // maximum number of fields/columns
-  private final int maxField;
-
-  /**
-   * We initialize and add the repeated varchar vector to the record batch in this
-   * constructor. Perform some sanity checks if the selected columns are valid or not.
-   * @param outputMutator  Used to create/modify schema in the record batch
-   * @param columns  List of columns selected in the query
-   * @param isStarQuery  boolean to indicate if all fields are selected or not
-   * @throws SchemaChangeException
-   */
-  public RepeatedVarCharOutput(OutputMutator outputMutator, Collection<SchemaPath> columns, boolean isStarQuery) throws SchemaChangeException {
-    super();
-
-    MaterializedField field = MaterializedField.create(COL_NAME, Types.repeated(TypeProtos.MinorType.VARCHAR));
-    this.vector = outputMutator.addField(field, RepeatedVarCharVector.class);
-
-    this.mutator = vector.getMutator();
-
-
-    { // setup fields
-      List<Integer> columnIds = new ArrayList<>();
-      if (!isStarQuery) {
-        String pathStr;
-        for (SchemaPath path : columns) {
-          assert path.getRootSegment().isNamed() : "root segment should be named";
-          pathStr = path.getRootSegment().getPath();
-          Preconditions.checkArgument(COL_NAME.equalsIgnoreCase(pathStr) || (SchemaPath.DYNAMIC_STAR.equals(pathStr) && path.getRootSegment().getChild() == null),
-              String.format("Selected column '%s' must have name 'columns' or must be plain '*'", pathStr));
-
-          if (path.getRootSegment().getChild() != null) {
-            Preconditions.checkArgument(path.getRootSegment().getChild().isArray(),
-              String.format("Selected column '%s' must be an array index", pathStr));
-            int index = path.getRootSegment().getChild().getArraySegment().getIndex();
-            columnIds.add(index);
-          }
-        }
-        Collections.sort(columnIds);
-
-      }
-
-      boolean[] fields = new boolean[MAXIMUM_NUMBER_COLUMNS];
-
-      int maxField = fields.length;
-
-      if(isStarQuery){
-        Arrays.fill(fields, true);
-      }else{
-        for(Integer i : columnIds){
-          maxField = 0;
-          maxField = Math.max(maxField, i);
-          fields[i] = true;
-        }
-      }
-      this.collectedFields = fields;
-      this.maxField = maxField;
-    }
-
-
-  }
-
-  /**
-   * Start a new record batch. Resets all the offsets and pointers that
-   * store buffer addresses
-   */
-  @Override
-  public void startBatch() {
-    this.recordStart = characterDataOriginal;
-    this.fieldOpen = false;
-    this.batchIndex = 0;
-    this.fieldIndex = -1;
-    this.collect = true;
-
-    loadRepeatedOffsetAddress();
-    loadVarCharOffsetAddress();
-    loadVarCharDataAddress();
-  }
-
-  private void loadRepeatedOffsetAddress(){
-    DrillBuf buf = vector.getOffsetVector().getBuffer();
-    checkBuf(buf);
-    this.repeatedOffset = buf.memoryAddress() + 4;
-    this.repeatedOffsetOriginal = buf.memoryAddress() + 4;
-    this.repeatedOffsetMax = buf.memoryAddress() + buf.capacity();
-  }
-
-  private void loadVarCharDataAddress(){
-    DrillBuf buf = vector.getDataVector().getBuffer();
-    checkBuf(buf);
-    this.characterData = buf.memoryAddress();
-    this.characterDataOriginal = buf.memoryAddress();
-    this.characterDataMax = buf.memoryAddress() + buf.capacity();
-  }
-
-  private void loadVarCharOffsetAddress(){
-    DrillBuf buf = vector.getDataVector().getOffsetVector().getBuffer();
-    checkBuf(buf);
-    this.charLengthOffset = buf.memoryAddress() + 4;
-    this.charLengthOffsetOriginal = buf.memoryAddress() + 4; // add four as offsets conceptually start at 1. (first item is 0..1)
-    this.charLengthOffsetMax = buf.memoryAddress() + buf.capacity();
-  }
-
-  private void expandVarCharOffsets(){
-    vector.getDataVector().getOffsetVector().reAlloc();
-    long diff = charLengthOffset - charLengthOffsetOriginal;
-    loadVarCharOffsetAddress();
-    charLengthOffset += diff;
-  }
-
-  private void expandVarCharData(){
-    vector.getDataVector().reAlloc();
-    long diff = characterData - characterDataOriginal;
-    loadVarCharDataAddress();
-    characterData += diff;
-  }
-
-  private void expandRepeatedOffsets(){
-    vector.getOffsetVector().reAlloc();
-    long diff = repeatedOffset - repeatedOffsetOriginal;
-    loadRepeatedOffsetAddress();
-    repeatedOffset += diff;
-  }
-
-  /**
-   * Helper method to check if the buffer we are accessing
-   * has a minimum reference count and has not been deallocated
-   * @param b  working drill buffer
-   */
-  private void checkBuf(DrillBuf b){
-    if(b.refCnt() < 1){
-      throw new IllegalStateException("Cannot access a dereferenced buffer.");
-    }
-  }
-
-  @Override
-  public void startField(int index) {
-    fieldIndex = index;
-    collect = collectedFields[index];
-    fieldOpen = true;
-  }
-
-  @Override
-  public boolean endField() {
-    fieldOpen = false;
-
-    if(charLengthOffset >= charLengthOffsetMax){
-      expandVarCharOffsets();
-    }
-
-    int newOffset = (int) (characterData - characterDataOriginal);
-    PlatformDependent.putInt(charLengthOffset, newOffset);
-    charLengthOffset += 4;
-    return fieldIndex < maxField;
-  }
-
-  @Override
-  public boolean endEmptyField() {
-    return endField();
-  }
-
-  @Override
-  public void append(byte data) {
-    if(!collect){
-      return;
-    }
-
-    if(characterData >= characterDataMax){
-      expandVarCharData();
-    }
-
-    PlatformDependent.putByte(characterData, data);
-    characterData++;
-
-  }
-
-  @Override
-  public long getRecordCount() {
-    return recordCount;
-  }
-
-  @Override
-  public boolean rowHasData() {
-    return this.recordStart < characterData;
-  }
-
-  @Override
-  public void finishRecord() {
-    this.recordStart = characterData;
-
-    if(fieldOpen){
-      endField();
-    }
-
-    if(repeatedOffset >= repeatedOffsetMax){
-      expandRepeatedOffsets();
-    }
-
-    int newOffset = ((int) (charLengthOffset - charLengthOffsetOriginal))/4;
-    PlatformDependent.putInt(repeatedOffset, newOffset);
-    repeatedOffset += 4;
-
-    // if there were no defined fields, skip.
-    if(fieldIndex > -1){
-      batchIndex++;
-      recordCount++;
-    }
-
-
-  }
-
-  /**
-   * This method is a helper method added for DRILL-951
-   * TextRecordReader to call this method to get field names out
-   * @return array of field data strings
-   */
-  public String [] getTextOutput () throws ExecutionSetupException {
-    if (recordCount == 0 || fieldIndex == -1) {
-      return null;
-    }
-
-    if (this.recordStart != characterData) {
-      throw new ExecutionSetupException("record text was requested before finishing record");
-    }
-
-    //Currently only first line header is supported. Return only first record.
-    int retSize = fieldIndex+1;
-    String [] out = new String [retSize];
-
-    RepeatedVarCharVector.Accessor a = this.vector.getAccessor();
-    for (int i=0; i<retSize; i++){
-      out[i] = a.getSingleObject(0,i).toString();
-    }
-    return out;
-  }
-
-  @Override
-  public void finishBatch() { }
-}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextInput.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextInput.java
deleted file mode 100644
index 6cf554d..0000000
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextInput.java
+++ /dev/null
@@ -1,365 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.drill.exec.store.easy.text.compliant;
-
-import io.netty.buffer.DrillBuf;
-import io.netty.util.internal.PlatformDependent;
-
-import java.io.EOFException;
-import java.io.IOException;
-import java.io.InputStream;
-import java.nio.ByteBuffer;
-
-import org.apache.hadoop.fs.ByteBufferReadable;
-import org.apache.hadoop.fs.FSDataInputStream;
-import org.apache.hadoop.fs.Seekable;
-import org.apache.hadoop.io.compress.CompressionInputStream;
-
-import org.apache.drill.shaded.guava.com.google.common.base.Preconditions;
-
-import static org.apache.drill.exec.memory.BoundsChecking.rangeCheck;
-
-/**
- * Class that fronts an InputStream to provide a byte consumption interface.
- * Also manages only reading lines to and from each split.
- */
-final class TextInput {
-
-  private final byte[] lineSeparator;
-  private final byte normalizedLineSeparator;
-  private final TextParsingSettings settings;
-
-  private long lineCount;
-  private long charCount;
-
-  /**
-   * The starting position in the file.
-   */
-  private final long startPos;
-  private final long endPos;
-
-  private int bufferMark;
-  private long streamMark;
-
-  private long streamPos;
-
-  private final Seekable seekable;
-  private final FSDataInputStream inputFS;
-  private final InputStream input;
-
-  private final DrillBuf buffer;
-  private final ByteBuffer underlyingBuffer;
-  private final long bStart;
-  private final long bStartMinus1;
-
-  private final boolean bufferReadable;
-
-  /**
-   * Whether there was a possible partial line separator on the previous
-   * read so we dropped it and it should be appended to next read.
-   */
-  private int remByte = -1;
-
-  /**
-   * The current position in the buffer.
-   */
-  public int bufferPtr;
-
-  /**
-   * The quantity of valid data in the buffer.
-   */
-  public int length = -1;
-
-  private boolean endFound = false;
-
-  /**
-   * Creates a new instance with the mandatory characters for handling newlines transparently.
-   * lineSeparator the sequence of characters that represent a newline, as defined in {@link Format#getLineSeparator()}
-   * normalizedLineSeparator the normalized newline character (as defined in {@link Format#getNormalizedNewline()}) that is used to replace any lineSeparator sequence found in the input.
-   */
-  public TextInput(TextParsingSettings settings, InputStream input, DrillBuf readBuffer, long startPos, long endPos) {
-    this.lineSeparator = settings.getNewLineDelimiter();
-    byte normalizedLineSeparator = settings.getNormalizedNewLine();
-    Preconditions.checkArgument(input instanceof Seekable, "Text input only supports an InputStream that supports Seekable.");
-    boolean isCompressed = input instanceof CompressionInputStream;
-    Preconditions.checkArgument(!isCompressed || startPos == 0, "Cannot use split on compressed stream.");
-
-    // splits aren't allowed with compressed data.  The split length will be the compressed size which means we'll normally end prematurely.
-    if(isCompressed && endPos > 0){
-      endPos = Long.MAX_VALUE;
-    }
-
-    this.input = input;
-    this.seekable = (Seekable) input;
-    this.settings = settings;
-
-    if(input instanceof FSDataInputStream){
-      this.inputFS = (FSDataInputStream) input;
-      this.bufferReadable = inputFS.getWrappedStream() instanceof ByteBufferReadable;
-    }else{
-      this.inputFS = null;
-      this.bufferReadable = false;
-    }
-
-    this.startPos = startPos;
-    this.endPos = endPos;
-
-    this.normalizedLineSeparator = normalizedLineSeparator;
-
-    this.buffer = readBuffer;
-    this.bStart = buffer.memoryAddress();
-    this.bStartMinus1 = bStart -1;
-    this.underlyingBuffer = buffer.nioBuffer(0, buffer.capacity());
-  }
-
-  /**
-   * Test the input to position for read start.  If the input is a non-zero split or
-   * splitFirstLine is enabled, input will move to appropriate complete line.
-   * @throws IOException
-   */
-  final void start() throws IOException {
-    lineCount = 0;
-    if(startPos > 0){
-      seekable.seek(startPos);
-    }
-
-    updateBuffer();
-    if (length > 0) {
-      if(startPos > 0 || settings.isSkipFirstLine()){
-
-        // move to next full record.
-        skipLines(1);
-      }
-    }
-  }
-
-
-  /**
-   * Helper method to get the most recent characters consumed since the last record started.
-   * May get an incomplete string since we don't support stream rewind.  Returns empty string for now.
-   * @return String of last few bytes.
-   * @throws IOException
-   */
-  public String getStringSinceMarkForError() throws IOException {
-    return " ";
-  }
-
-  long getPos(){
-    return streamPos + bufferPtr;
-  }
-
-  public void mark(){
-    streamMark = streamPos;
-    bufferMark = bufferPtr;
-  }
-
-  /**
-   * read some more bytes from the stream.  Uses the zero copy interface if available.  Otherwise, does byte copy.
-   * @throws IOException
-   */
-  private void read() throws IOException {
-    if(bufferReadable){
-
-      if(remByte != -1){
-        for (int i = 0; i <= remByte; i++) {
-          underlyingBuffer.put(lineSeparator[i]);
-        }
-        remByte = -1;
-      }
-      length = inputFS.read(underlyingBuffer);
-
-    }else{
-
-      byte[] b = new byte[underlyingBuffer.capacity()];
-      if(remByte != -1){
-        int remBytesNum = remByte + 1;
-        System.arraycopy(lineSeparator, 0, b, 0, remBytesNum);
-        length = input.read(b, remBytesNum, b.length - remBytesNum);
-        remByte = -1;
-      }else{
-        length = input.read(b);
-      }
-      underlyingBuffer.put(b);
-    }
-  }
-
-
-  /**
-   * Read more data into the buffer.  Will also manage split end conditions.
-   * @throws IOException
-   */
-  private void updateBuffer() throws IOException {
-    streamPos = seekable.getPos();
-    underlyingBuffer.clear();
-
-    if(endFound){
-      length = -1;
-      return;
-    }
-
-    read();
-
-    // check our data read allowance.
-    if(streamPos + length >= this.endPos){
-      updateLengthBasedOnConstraint();
-    }
-
-    charCount += bufferPtr;
-    bufferPtr = 1;
-
-    buffer.writerIndex(underlyingBuffer.limit());
-    buffer.readerIndex(underlyingBuffer.position());
-
-  }
-
-  /**
-   * Checks to see if we can go over the end of our bytes constraint on the data.  If so,
-   * adjusts so that we can only read to the last character of the first line that crosses
-   * the split boundary.
-   */
-  private void updateLengthBasedOnConstraint() {
-    final long max = bStart + length;
-    for(long m = bStart + (endPos - streamPos); m < max; m++) {
-      for (int i = 0; i < lineSeparator.length; i++) {
-        long mPlus = m + i;
-        if (mPlus < max) {
-          // we found a line separator and don't need to consult the next byte.
-          if (lineSeparator[i] == PlatformDependent.getByte(mPlus) && i == lineSeparator.length - 1) {
-            length = (int) (mPlus - bStart) + 1;
-            endFound = true;
-            return;
-          }
-        } else {
-          // the last N characters of the read were remnant bytes. We'll hold off on dealing with these bytes until the next read.
-          remByte = i;
-          length = length - i;
-          return;
-        }
-      }
-    }
-  }
-
-  /**
-   * Get next byte from stream.  Also maintains the current line count.  Will throw a StreamFinishedPseudoException
-   * when the stream has run out of bytes.
-   * @return next byte from stream.
-   * @throws IOException
-   */
-  public final byte nextChar() throws IOException {
-    byte byteChar = nextCharNoNewLineCheck();
-    int bufferPtrTemp = bufferPtr - 1;
-    if (byteChar == lineSeparator[0]) {
-       for (int i = 1; i < lineSeparator.length; i++, bufferPtrTemp++) {
-         if (lineSeparator[i] != buffer.getByte(bufferPtrTemp)) {
-           return byteChar;
-         }
-       }
-
-        lineCount++;
-        byteChar = normalizedLineSeparator;
-
-        // we don't need to update buffer position if line separator is one byte long
-        if (lineSeparator.length > 1) {
-          bufferPtr += (lineSeparator.length - 1);
-          if (bufferPtr >= length) {
-            if (length != -1) {
-              updateBuffer();
-            } else {
-              throw StreamFinishedPseudoException.INSTANCE;
-            }
-          }
-        }
-      }
-
-    return byteChar;
-  }
-
-  /**
-   * Get next byte from stream.  Do no maintain any line count  Will throw a StreamFinishedPseudoException
-   * when the stream has run out of bytes.
-   * @return next byte from stream.
-   * @throws IOException
-   */
-  public final byte nextCharNoNewLineCheck() throws IOException {
-
-    if (length == -1) {
-      throw StreamFinishedPseudoException.INSTANCE;
-    }
-
-    rangeCheck(buffer, bufferPtr - 1, bufferPtr);
-
-    byte byteChar = PlatformDependent.getByte(bStartMinus1 + bufferPtr);
-
-    if (bufferPtr >= length) {
-      if (length != -1) {
-        updateBuffer();
-        bufferPtr--;
-      } else {
-        throw StreamFinishedPseudoException.INSTANCE;
-      }
-    }
-
-    bufferPtr++;
-
-    return byteChar;
-  }
-
-  /**
-   * Number of lines read since the start of this split.
-   * @return
-   */
-  public final long lineCount() {
-    return lineCount;
-  }
-
-  /**
-   * Skip forward the number of line delimiters.  If you are in the middle of a line,
-   * a value of 1 will skip to the start of the next record.
-   * @param lines Number of lines to skip.
-   * @throws IOException
-   */
-  public final void skipLines(int lines) throws IOException {
-    if (lines < 1) {
-      return;
-    }
-    long expectedLineCount = this.lineCount + lines;
-
-    try {
-      do {
-        nextChar();
-      } while (lineCount < expectedLineCount);
-      if (lineCount < lines) {
-        throw new IllegalArgumentException("Unable to skip " + lines + " lines from line " + (expectedLineCount - lines) + ". End of input reached");
-      }
-    } catch (EOFException ex) {
-      throw new IllegalArgumentException("Unable to skip " + lines + " lines from line " + (expectedLineCount - lines) + ". End of input reached");
-    }
-  }
-
-  public final long charCount() {
-    return charCount + bufferPtr;
-  }
-
-  public long getLineCount() {
-    return lineCount;
-  }
-
-  public void close() throws IOException{
-    input.close();
-  }
-}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextOutput.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextOutput.java
deleted file mode 100644
index 1042994..0000000
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextOutput.java
+++ /dev/null
@@ -1,87 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.drill.exec.store.easy.text.compliant;
-
-/* Base class for producing output record batches while dealing with
- * Text files.
- */
-abstract class TextOutput {
-  static final org.slf4j.Logger logger = org.slf4j.LoggerFactory.getLogger(TextOutput.class);
-
-  /**
-   * Start processing a new field within a record.
-   * @param index  index within the record
-   */
-  public abstract void startField(int index);
-
-  /**
-   * End processing a field within a record.
-   * @return  true if engine should continue processing record.  false if rest of record can be skipped.
-   */
-  public abstract boolean endField();
-
-  /**
-   * Shortcut that lets the output know that we are closing ending a field with no data.
-   * @return true if engine should continue processing record.  false if rest of record can be skipped.
-   */
-  public abstract boolean endEmptyField();
-
-  /**
-   * Add the provided data but drop any whitespace.
-   * @param data
-   */
-  public void appendIgnoringWhitespace(byte data){
-    if(TextReader.isWhite(data)){
-      // noop
-    }else{
-      append(data);
-    }
-  }
-
-  /**
-   * This function appends the byte to the output character data buffer
-   * @param data  current byte read
-   */
-  public abstract void append(byte data);
-
-  /**
-   * Completes the processing of a given record. Also completes the processing of the
-   * last field being read.
-   */
-  public abstract void finishRecord();
-
-  /**
-   *  Return the total number of records (across batches) processed
-   */
-  public abstract long getRecordCount();
-
-  /**
-   * Informs output to setup for new record batch.
-   */
-  public abstract void startBatch();
-
-  /**
-   * Does any final cleanup that is required for closing a batch.  Example might include closing the last field.
-   */
-  public abstract void finishBatch();
-
-  /**
-   * Helper method to check if the current record has any non-empty fields
-   */
-  public abstract boolean rowHasData();
-}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextParsingContext.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextParsingContext.java
deleted file mode 100644
index 684c453..0000000
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextParsingContext.java
+++ /dev/null
@@ -1,122 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.drill.exec.store.easy.text.compliant;
-
-import java.io.IOException;
-
-import com.univocity.parsers.common.ParsingContext;
-
-class TextParsingContext implements ParsingContext {
-
-  private final TextInput input;
-  private final TextOutput output;
-  protected boolean stopped = false;
-
-  private int[] extractedIndexes = null;
-
-  public TextParsingContext(TextInput input, TextOutput output) {
-    this.input = input;
-    this.output = output;
-  }
-
-  /**
-   * {@inheritDoc}
-   */
-  @Override
-  public void stop() {
-    stopped = true;
-  }
-
-  /**
-   * {@inheritDoc}
-   */
-  @Override
-  public boolean isStopped() {
-    return stopped;
-  }
-
-  /**
-   * {@inheritDoc}
-   */
-  @Override
-  public long currentLine() {
-    return input.lineCount();
-  }
-
-  /**
-   * {@inheritDoc}
-   */
-  @Override
-  public long currentChar() {
-    return input.charCount();
-  }
-
-  /**
-   * {@inheritDoc}
-   */
-  @Override
-  public int currentColumn() {
-    return -1;
-  }
-
-  /**
-   * {@inheritDoc}
-   */
-  @Override
-  public String[] headers() {
-    return new String[]{};
-  }
-
-  /**
-   * {@inheritDoc}
-   */
-  @Override
-  public int[] extractedFieldIndexes() {
-    return extractedIndexes;
-  }
-
-  /**
-   * {@inheritDoc}
-   */
-  @Override
-  public long currentRecord() {
-    return output.getRecordCount();
-  }
-
-  /**
-   * {@inheritDoc}
-   */
-  @Override
-  public String currentParsedContent() {
-    try {
-      return input.getStringSinceMarkForError();
-    } catch (IOException e) {
-      throw new RuntimeException(e);
-    }
-  }
-
-  @Override
-  public void skipLines(int lines) {
-  }
-
-  @Override
-  public boolean columnsReordered() {
-    return false;
-  }
-}
-
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextParsingSettings.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextParsingSettings.java
deleted file mode 100644
index 36dd4f9..0000000
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextParsingSettings.java
+++ /dev/null
@@ -1,291 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.drill.exec.store.easy.text.compliant;
-
-import org.apache.drill.exec.store.easy.text.TextFormatPlugin.TextFormatConfig;
-
-import org.apache.drill.shaded.guava.com.google.common.base.Charsets;
-
-public class TextParsingSettings {
-  static final org.slf4j.Logger logger = org.slf4j.LoggerFactory.getLogger(TextParsingSettings.class);
-
-  public static final TextParsingSettings DEFAULT = new TextParsingSettings();
-
-  private String emptyValue = null;
-  private boolean parseUnescapedQuotes = true;
-  private byte quote = b('"');
-  private byte quoteEscape = b('"');
-  private byte delimiter = b(',');
-  private byte comment = b('#');
-
-  private long maxCharsPerColumn = Character.MAX_VALUE;
-  private byte normalizedNewLine = b('\n');
-  private byte[] newLineDelimiter = {normalizedNewLine};
-  private boolean ignoreLeadingWhitespaces = false;
-  private boolean ignoreTrailingWhitespaces = false;
-  private String lineSeparatorString = "\n";
-  private boolean skipFirstLine = false;
-
-  private boolean headerExtractionEnabled = false;
-  private boolean useRepeatedVarChar = true;
-  private int numberOfRecordsToRead = -1;
-
-  public void set(TextFormatConfig config){
-    this.quote = bSafe(config.getQuote(), "quote");
-    this.quoteEscape = bSafe(config.getEscape(), "escape");
-    this.newLineDelimiter = config.getLineDelimiter().getBytes(Charsets.UTF_8);
-    this.delimiter = bSafe(config.getFieldDelimiter(), "fieldDelimiter");
-    this.comment = bSafe(config.getComment(), "comment");
-    this.skipFirstLine = config.isSkipFirstLine();
-    this.headerExtractionEnabled = config.isHeaderExtractionEnabled();
-    if (this.headerExtractionEnabled) {
-      // In case of header TextRecordReader will use set of VarChar vectors vs RepeatedVarChar
-      this.useRepeatedVarChar = false;
-    }
-  }
-
-  public byte getComment(){
-    return comment;
-  }
-
-  public boolean isSkipFirstLine() {
-    return skipFirstLine;
-  }
-
-  public void setSkipFirstLine(boolean skipFirstLine) {
-    this.skipFirstLine = skipFirstLine;
-  }
-
-  public boolean isUseRepeatedVarChar() {
-    return useRepeatedVarChar;
-  }
-
-  public void setUseRepeatedVarChar(boolean useRepeatedVarChar) {
-    this.useRepeatedVarChar = useRepeatedVarChar;
-  }
-
-
-  private static byte bSafe(char c, String name){
-    if(c > Byte.MAX_VALUE) {
-      throw new IllegalArgumentException(String.format("Failure validating configuration option %s.  Expected a "
-          + "character between 0 and 127 but value was actually %d.", name, (int) c));
-    }
-    return (byte) c;
-  }
-
-  private static byte b(char c){
-    return (byte) c;
-  }
-
-  public byte[] getNewLineDelimiter() {
-    return newLineDelimiter;
-  }
-
-  /**
-   * Returns the character used for escaping values where the field delimiter is part of the value. Defaults to '"'
-   * @return the quote character
-   */
-  public byte getQuote() {
-    return quote;
-  }
-
-  /**
-   * Defines the character used for escaping values where the field delimiter is part of the value. Defaults to '"'
-   * @param quote the quote character
-   */
-  public void setQuote(byte quote) {
-    this.quote = quote;
-  }
-
-  public String getLineSeparatorString(){
-    return lineSeparatorString;
-  }
-
-
-  /**
-   * Identifies whether or not a given character is used for escaping values where the field delimiter is part of the value
-   * @param ch the character to be verified
-   * @return true if the given character is the character used for escaping values, false otherwise
-   */
-  public boolean isQuote(byte ch) {
-    return this.quote == ch;
-  }
-
-  /**
-   * Returns the character used for escaping quotes inside an already quoted value. Defaults to '"'
-   * @return the quote escape character
-   */
-  public byte getQuoteEscape() {
-    return quoteEscape;
-  }
-
-  /**
-   * Defines the character used for escaping quotes inside an already quoted value. Defaults to '"'
-   * @param quoteEscape the quote escape character
-   */
-  public void setQuoteEscape(byte quoteEscape) {
-    this.quoteEscape = quoteEscape;
-  }
-
-  /**
-   * Identifies whether or not a given character is used for escaping quotes inside an already quoted value.
-   * @param ch the character to be verified
-   * @return true if the given character is the quote escape character, false otherwise
-   */
-  public boolean isQuoteEscape(byte ch) {
-    return this.quoteEscape == ch;
-  }
-
-  /**
-   * Returns the field delimiter character. Defaults to ','
-   * @return the field delimiter character
-   */
-  public byte getDelimiter() {
-    return delimiter;
-  }
-
-  /**
-   * Defines the field delimiter character. Defaults to ','
-   * @param delimiter the field delimiter character
-   */
-  public void setDelimiter(byte delimiter) {
-    this.delimiter = delimiter;
-  }
-
-  /**
-   * Identifies whether or not a given character represents a field delimiter
-   * @param ch the character to be verified
-   * @return true if the given character is the field delimiter character, false otherwise
-   */
-  public boolean isDelimiter(byte ch) {
-    return this.delimiter == ch;
-  }
-
-  /**
-   * Returns the String representation of an empty value (defaults to null)
-   *
-   * <p>When reading, if the parser does not read any character from the input, and the input is within quotes, the empty is used instead of an empty string
-   *
-   * @return the String representation of an empty value
-   */
-  public String getEmptyValue() {
-    return emptyValue;
-  }
-
-  /**
-   * Sets the String representation of an empty value (defaults to null)
-   *
-   * <p>When reading, if the parser does not read any character from the input, and the input is within quotes, the empty is used instead of an empty string
-   *
-   * @param emptyValue the String representation of an empty value
-   */
-  public void setEmptyValue(String emptyValue) {
-    this.emptyValue = emptyValue;
-  }
-
-
-  /**
-   * Indicates whether the CSV parser should accept unescaped quotes inside quoted values and parse them normally. Defaults to {@code true}.
-   * @return a flag indicating whether or not the CSV parser should accept unescaped quotes inside quoted values.
-   */
-  public boolean isParseUnescapedQuotes() {
-    return parseUnescapedQuotes;
-  }
-
-  /**
-   * Configures how to handle unescaped quotes inside quoted values. If set to {@code true}, the parser will parse the quote normally as part of the value.
-   * If set the {@code false}, a {@link com.univocity.parsers.common.TextParsingException} will be thrown. Defaults to {@code true}.
-   * @param parseUnescapedQuotes indicates whether or not the CSV parser should accept unescaped quotes inside quoted values.
-   */
-  public void setParseUnescapedQuotes(boolean parseUnescapedQuotes) {
-    this.parseUnescapedQuotes = parseUnescapedQuotes;
-  }
-
-  /**
-   * Indicates whether or not the first valid record parsed from the input should be considered as the row containing the names of each column
-   * @return true if the first valid record parsed from the input should be considered as the row containing the names of each column, false otherwise
-   */
-  public boolean isHeaderExtractionEnabled() {
-    return headerExtractionEnabled;
-  }
-
-  /**
-   * Defines whether or not the first valid record parsed from the input should be considered as the row containing the names of each column
-   * @param headerExtractionEnabled a flag indicating whether the first valid record parsed from the input should be considered as the row containing the names of each column
-   */
-  public void setHeaderExtractionEnabled(boolean headerExtractionEnabled) {
-    this.headerExtractionEnabled = headerExtractionEnabled;
-  }
-
-  /**
-   * The number of valid records to be parsed before the process is stopped. A negative value indicates there's no limit (defaults to -1).
-   * @return the number of records to read before stopping the parsing process.
-   */
-  public int getNumberOfRecordsToRead() {
-    return numberOfRecordsToRead;
-  }
-
-  /**
-   * Defines the number of valid records to be parsed before the process is stopped. A negative value indicates there's no limit (defaults to -1).
-   * @param numberOfRecordsToRead the number of records to read before stopping the parsing process.
-   */
-  public void setNumberOfRecordsToRead(int numberOfRecordsToRead) {
-    this.numberOfRecordsToRead = numberOfRecordsToRead;
-  }
-
-  public long getMaxCharsPerColumn() {
-    return maxCharsPerColumn;
-  }
-
-  public void setMaxCharsPerColumn(long maxCharsPerColumn) {
-    this.maxCharsPerColumn = maxCharsPerColumn;
-  }
-
-  public void setComment(byte comment) {
-    this.comment = comment;
-  }
-
-  public byte getNormalizedNewLine() {
-    return normalizedNewLine;
-  }
-
-  public void setNormalizedNewLine(byte normalizedNewLine) {
-    this.normalizedNewLine = normalizedNewLine;
-  }
-
-  public boolean isIgnoreLeadingWhitespaces() {
-    return ignoreLeadingWhitespaces;
-  }
-
-  public void setIgnoreLeadingWhitespaces(boolean ignoreLeadingWhitespaces) {
-    this.ignoreLeadingWhitespaces = ignoreLeadingWhitespaces;
-  }
-
-  public boolean isIgnoreTrailingWhitespaces() {
-    return ignoreTrailingWhitespaces;
-  }
-
-  public void setIgnoreTrailingWhitespaces(boolean ignoreTrailingWhitespaces) {
-    this.ignoreTrailingWhitespaces = ignoreTrailingWhitespaces;
-  }
-
-
-
-
-
-}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextReader.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextReader.java
deleted file mode 100644
index 7a9ed46..0000000
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextReader.java
+++ /dev/null
@@ -1,511 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.drill.exec.store.easy.text.compliant;
-
-import io.netty.buffer.DrillBuf;
-
-import java.io.IOException;
-
-import org.apache.drill.common.exceptions.UserException;
-
-import com.univocity.parsers.common.TextParsingException;
-
-/*******************************************************************************
- * Portions Copyright 2014 uniVocity Software Pty Ltd
- ******************************************************************************/
-
-/**
- * A byte-based Text parser implementation. Builds heavily upon the uniVocity parsers. Customized for UTF8 parsing and
- * DrillBuf support.
- */
-final class TextReader {
-  static final org.slf4j.Logger logger = org.slf4j.LoggerFactory.getLogger(TextReader.class);
-
-  private static final byte NULL_BYTE = (byte) '\0';
-
-  private final TextParsingContext context;
-
-  private final long recordsToRead;
-  private final TextParsingSettings settings;
-
-  private final TextInput input;
-  private final TextOutput output;
-  private final DrillBuf workBuf;
-
-  private byte ch;
-
-  // index of the field within this record
-  private int fieldIndex;
-
-  /** Behavior settings **/
-  private final boolean ignoreTrailingWhitespace;
-  private final boolean ignoreLeadingWhitespace;
-  private final boolean parseUnescapedQuotes;
-
-  /** Key Characters **/
-  private final byte comment;
-  private final byte delimiter;
-  private final byte quote;
-  private final byte quoteEscape;
-  private final byte newLine;
-
-  /**
-   * The CsvParser supports all settings provided by {@link CsvParserSettings}, and requires this configuration to be
-   * properly initialized.
-   * @param settings  the parser configuration
-   * @param input  input stream
-   * @param output  interface to produce output record batch
-   * @param workBuf  working buffer to handle whitespaces
-   */
-  public TextReader(TextParsingSettings settings, TextInput input, TextOutput output, DrillBuf workBuf) {
-    this.context = new TextParsingContext(input, output);
-    this.workBuf = workBuf;
-    this.settings = settings;
-
-    this.recordsToRead = settings.getNumberOfRecordsToRead() == -1 ? Long.MAX_VALUE : settings.getNumberOfRecordsToRead();
-
-    this.ignoreTrailingWhitespace = settings.isIgnoreTrailingWhitespaces();
-    this.ignoreLeadingWhitespace = settings.isIgnoreLeadingWhitespaces();
-    this.parseUnescapedQuotes = settings.isParseUnescapedQuotes();
-    this.delimiter = settings.getDelimiter();
-    this.quote = settings.getQuote();
-    this.quoteEscape = settings.getQuoteEscape();
-    this.newLine = settings.getNormalizedNewLine();
-    this.comment = settings.getComment();
-
-    this.input = input;
-    this.output = output;
-
-  }
-
-  public TextOutput getOutput(){
-    return output;
-  }
-
-  /* Check if the given byte is a white space. As per the univocity text reader
-   * any ASCII <= ' ' is considered a white space. However since byte in JAVA is signed
-   * we have an additional check to make sure its not negative
-   */
-  static final boolean isWhite(byte b){
-    return b <= ' ' && b > -1;
-  }
-
-  // Inform the output interface to indicate we are starting a new record batch
-  public void resetForNextBatch(){
-    output.startBatch();
-  }
-
-  public long getPos(){
-    return input.getPos();
-  }
-
-  /**
-   * Function encapsulates parsing an entire record, delegates parsing of the
-   * fields to parseField() function.
-   * We mark the start of the record and if there are any failures encountered (OOM for eg)
-   * then we reset the input stream to the marked position
-   * @return  true if parsing this record was successful; false otherwise
-   * @throws IOException
-   */
-  private boolean parseRecord() throws IOException {
-    final byte newLine = this.newLine;
-    final TextInput input = this.input;
-
-    input.mark();
-
-    fieldIndex = 0;
-    if (isWhite(ch) && ignoreLeadingWhitespace) {
-      skipWhitespace();
-    }
-
-    int fieldsWritten = 0;
-    try{
-      boolean earlyTerm = false;
-      while (ch != newLine) {
-        earlyTerm = !parseField();
-        fieldsWritten++;
-        if (ch != newLine) {
-          ch = input.nextChar();
-          if (ch == newLine) {
-            output.startField(fieldsWritten++);
-            output.endEmptyField();
-            break;
-          }
-        }
-        if(earlyTerm){
-          if(ch != newLine){
-            input.skipLines(1);
-          }
-          break;
-        }
-      }
-    }catch(StreamFinishedPseudoException e){
-      // if we've written part of a field or all of a field, we should send this row.
-      if(fieldsWritten == 0 && !output.rowHasData()){
-        throw e;
-      }
-    }
-
-    output.finishRecord();
-    return true;
-  }
-
-  /**
-   * Function parses an individual field and ignores any white spaces encountered
-   * by not appending it to the output vector
-   * @throws IOException
-   */
-  private void parseValueIgnore() throws IOException {
-    final byte newLine = this.newLine;
-    final byte delimiter = this.delimiter;
-    final TextOutput output = this.output;
-    final TextInput input = this.input;
-
-    byte ch = this.ch;
-    while (ch != delimiter && ch != newLine) {
-      output.appendIgnoringWhitespace(ch);
-//      fieldSize++;
-      ch = input.nextChar();
-    }
-    this.ch = ch;
-  }
-
-  /**
-   * Function parses an individual field and appends all characters till the delimeter (or newline)
-   * to the output, including white spaces
-   * @throws IOException
-   */
-  private void parseValueAll() throws IOException {
-    final byte newLine = this.newLine;
-    final byte delimiter = this.delimiter;
-    final TextOutput output = this.output;
-    final TextInput input = this.input;
-
-    byte ch = this.ch;
-    while (ch != delimiter && ch != newLine) {
-      output.append(ch);
-      ch = input.nextChar();
-    }
-    this.ch = ch;
-  }
-
-  /**
-   * Function simply delegates the parsing of a single field to the actual implementation based on parsing config
-   * @throws IOException
-   */
-  private void parseValue() throws IOException {
-    if (ignoreTrailingWhitespace) {
-      parseValueIgnore();
-    }else{
-      parseValueAll();
-    }
-  }
-
-  /**
-   * Recursive function invoked when a quote is encountered. Function also
-   * handles the case when there are non-white space characters in the field
-   * after the quoted value.
-   * @param prev  previous byte read
-   * @throws IOException
-   */
-  private void parseQuotedValue(byte prev) throws IOException {
-    final byte newLine = this.newLine;
-    final byte delimiter = this.delimiter;
-    final TextOutput output = this.output;
-    final TextInput input = this.input;
-    final byte quote = this.quote;
-
-    ch = input.nextCharNoNewLineCheck();
-
-    while (!(prev == quote && (ch == delimiter || ch == newLine || isWhite(ch)))) {
-      if (ch != quote) {
-        if (prev == quote) { // unescaped quote detected
-          if (parseUnescapedQuotes) {
-            output.append(quote);
-            output.append(ch);
-            parseQuotedValue(ch);
-            break;
-          } else {
-            throw new TextParsingException(
-                context,
-                "Unescaped quote character '"
-                    + quote
-                    + "' inside quoted value of CSV field. To allow unescaped quotes, set 'parseUnescapedQuotes' to 'true' in the CSV parser settings. Cannot parse CSV input.");
-          }
-        }
-        output.append(ch);
-        prev = ch;
-      } else if (prev == quoteEscape) {
-        output.append(quote);
-        prev = NULL_BYTE;
-      } else {
-        prev = ch;
-      }
-      ch = input.nextCharNoNewLineCheck();
-    }
-
-    // Handles whitespaces after quoted value:
-    // Whitespaces are ignored (i.e., ch <= ' ') if they are not used as delimiters (i.e., ch != ' ')
-    // For example, in tab-separated files (TSV files), '\t' is used as delimiter and should not be ignored
-    // Content after whitespaces may be parsed if 'parseUnescapedQuotes' is enabled.
-    if (ch != newLine && ch <= ' ' && ch != delimiter) {
-      final DrillBuf workBuf = this.workBuf;
-      workBuf.resetWriterIndex();
-      do {
-        // saves whitespaces after value
-        workBuf.writeByte(ch);
-        ch = input.nextChar();
-        // found a new line, go to next record.
-        if (ch == newLine) {
-          return;
-        }
-      } while (ch <= ' ' && ch != delimiter);
-
-      // there's more stuff after the quoted value, not only empty spaces.
-      if (!(ch == delimiter || ch == newLine) && parseUnescapedQuotes) {
-
-        output.append(quote);
-        for(int i =0; i < workBuf.writerIndex(); i++){
-          output.append(workBuf.getByte(i));
-        }
-        // the next character is not the escape character, put it there
-        if (ch != quoteEscape) {
-          output.append(ch);
-        }
-        // sets this character as the previous character (may be escaping)
-        // calls recursively to keep parsing potentially quoted content
-        parseQuotedValue(ch);
-      }
-    }
-
-    if (!(ch == delimiter || ch == newLine)) {
-      throw new TextParsingException(context, "Unexpected character '" + ch
-          + "' following quoted value of CSV field. Expecting '" + delimiter + "'. Cannot parse CSV input.");
-    }
-  }
-
-  /**
-   * Captures the entirety of parsing a single field and based on the input delegates to the appropriate function
-   * @return
-   * @throws IOException
-   */
-  private final boolean parseField() throws IOException {
-
-    output.startField(fieldIndex++);
-
-    if (isWhite(ch) && ignoreLeadingWhitespace) {
-      skipWhitespace();
-    }
-
-    if (ch == delimiter) {
-      return output.endEmptyField();
-    } else {
-      if (ch == quote) {
-        parseQuotedValue(NULL_BYTE);
-      } else {
-        parseValue();
-      }
-
-      return output.endField();
-    }
-
-  }
-
-  /**
-   * Helper function to skip white spaces occurring at the current input stream.
-   * @throws IOException
-   */
-  private void skipWhitespace() throws IOException {
-    final byte delimiter = this.delimiter;
-    final byte newLine = this.newLine;
-    final TextInput input = this.input;
-
-    while (isWhite(ch) && ch != delimiter && ch != newLine) {
-      ch = input.nextChar();
-    }
-  }
-
-  /**
-   * Starting point for the reader. Sets up the input interface.
-   * @throws IOException
-   */
-  public final void start() throws IOException {
-    context.stopped = false;
-    input.start();
-  }
-
-
-  /**
-   * Parses the next record from the input. Will skip the line if its a comment,
-   * this is required when the file contains headers
-   * @throws IOException
-   */
-  public final boolean parseNext() throws IOException {
-    try {
-      while (!context.stopped) {
-        ch = input.nextChar();
-        if (ch == comment) {
-          input.skipLines(1);
-          continue;
-        }
-        break;
-      }
-      final long initialLineNumber = input.lineCount();
-      boolean success = parseRecord();
-      if (initialLineNumber + 1 < input.lineCount()) {
-        throw new TextParsingException(context, "Cannot use newline character within quoted string");
-      }
-
-      if (success) {
-        if (recordsToRead > 0 && context.currentRecord() >= recordsToRead) {
-          context.stop();
-        }
-        return true;
-      } else {
-        return false;
-      }
-
-    } catch (UserException ex) {
-      stopParsing();
-      throw ex;
-    } catch (StreamFinishedPseudoException ex) {
-      stopParsing();
-      return false;
-    } catch (Exception ex) {
-      try {
-        throw handleException(ex);
-      } finally {
-        stopParsing();
-      }
-    }
-  }
-
-  private void stopParsing(){
-
-  }
-
-  private String displayLineSeparators(String str, boolean addNewLine) {
-    if (addNewLine) {
-      if (str.contains("\r\n")) {
-        str = str.replaceAll("\\r\\n", "[\\\\r\\\\n]\r\n\t");
-      } else if (str.contains("\n")) {
-        str = str.replaceAll("\\n", "[\\\\n]\n\t");
-      } else {
-        str = str.replaceAll("\\r", "[\\\\r]\r\t");
-      }
-    } else {
-      str = str.replaceAll("\\n", "\\\\n");
-      str = str.replaceAll("\\r", "\\\\r");
-    }
-    return str;
-  }
-
-  /**
-   * Helper method to handle exceptions caught while processing text files and generate better error messages associated with
-   * the exception.
-   * @param ex  Exception raised
-   * @return
-   * @throws IOException
-   */
-  private TextParsingException handleException(Exception ex) throws IOException {
-
-    if (ex instanceof TextParsingException) {
-      throw (TextParsingException) ex;
-    }
-
-    if (ex instanceof ArrayIndexOutOfBoundsException) {
-      ex = UserException
-          .dataReadError(ex)
-          .message(
-              "Drill failed to read your text file.  Drill supports up to %d columns in a text file.  Your file appears to have more than that.",
-              RepeatedVarCharOutput.MAXIMUM_NUMBER_COLUMNS)
-          .build(logger);
-    }
-
-    String message = null;
-    String tmp = input.getStringSinceMarkForError();
-    char[] chars = tmp.toCharArray();
-    if (chars != null) {
-      int length = chars.length;
-      if (length > settings.getMaxCharsPerColumn()) {
-        message = "Length of parsed input (" + length
-            + ") exceeds the maximum number of characters defined in your parser settings ("
-            + settings.getMaxCharsPerColumn() + "). ";
-      }
-
-      if (tmp.contains("\n") || tmp.contains("\r")) {
-        tmp = displayLineSeparators(tmp, true);
-        String lineSeparator = displayLineSeparators(settings.getLineSeparatorString(), false);
-        message += "\nIdentified line separator characters in the parsed content. This may be the cause of the error. The line separator in your parser settings is set to '"
-            + lineSeparator + "'. Parsed content:\n\t" + tmp;
-      }
-
-      int nullCharacterCount = 0;
-      // ensuring the StringBuilder won't grow over Integer.MAX_VALUE to avoid OutOfMemoryError
-      int maxLength = length > Integer.MAX_VALUE / 2 ? Integer.MAX_VALUE / 2 - 1 : length;
-      StringBuilder s = new StringBuilder(maxLength);
-      for (int i = 0; i < maxLength; i++) {
-        if (chars[i] == '\0') {
-          s.append('\\');
-          s.append('0');
-          nullCharacterCount++;
-        } else {
-          s.append(chars[i]);
-        }
-      }
-      tmp = s.toString();
-
-      if (nullCharacterCount > 0) {
-        message += "\nIdentified "
-            + nullCharacterCount
-            + " null characters ('\0') on parsed content. This may indicate the data is corrupt or its encoding is invalid. Parsed content:\n\t"
-            + tmp;
-      }
-
-    }
-
-    throw new TextParsingException(context, message, ex);
-  }
-
-  /**
-   * Finish the processing of a batch, indicates to the output
-   * interface to wrap up the batch
-   */
-  public void finishBatch(){
-    output.finishBatch();
-//    System.out.println(String.format("line %d, cnt %d", input.getLineCount(), output.getRecordCount()));
-  }
-
-  /**
-   * Invoked once there are no more records and we are done with the
-   * current record reader to clean up state.
-   * @throws IOException
-   */
-  public void close() throws IOException{
-    input.close();
-  }
-
-  @Override
-  public String toString() {
-    return "TextReader[Line=" + context.currentLine()
-        + ", Column=" + context.currentChar()
-        + ", Record=" + context.currentRecord()
-        + ", Byte pos=" + getPos()
-        + "]";
-  }
-}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/package-info.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/package-info.java
deleted file mode 100644
index 0ed3155..0000000
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/package-info.java
+++ /dev/null
@@ -1,27 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-/**
- * Original version of the "compliant" text reader. This is version 2 of
- * the text reader. This version is retained for temporary backward
- * compatibility as we productize the newer version 3 based on the
- * row set framework.
- * <p>
- * TODO: Remove the files in this package and move the files from the
- * "v3" sub-package here once the version 3 implementation stabilizes.
- */
-package org.apache.drill.exec.store.easy.text.compliant;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/StreamFinishedPseudoException.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/StreamFinishedPseudoException.java
deleted file mode 100644
index 70c43b7..0000000
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/StreamFinishedPseudoException.java
+++ /dev/null
@@ -1,29 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.drill.exec.store.easy.text.compliant.v3;
-
-class StreamFinishedPseudoException extends RuntimeException {
-
-  public static final StreamFinishedPseudoException INSTANCE = new StreamFinishedPseudoException();
-
-  private StreamFinishedPseudoException() {
-    super("", null, false, true);
-
-  }
-
-}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/BaseFieldOutput.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/BaseFieldOutput.java
similarity index 98%
rename from exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/BaseFieldOutput.java
rename to exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/BaseFieldOutput.java
index 5dd4284..fc17186 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/BaseFieldOutput.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/BaseFieldOutput.java
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.drill.exec.store.easy.text.compliant.v3;
+package org.apache.drill.exec.store.easy.text.reader;
 
 import org.apache.drill.exec.physical.rowSet.RowSetLoader;
 import org.apache.drill.exec.vector.accessor.ScalarWriter;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/CompliantTextBatchReader.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/CompliantTextBatchReader.java
similarity index 98%
rename from exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/CompliantTextBatchReader.java
rename to exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/CompliantTextBatchReader.java
index 19b3dbd..877d963 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/CompliantTextBatchReader.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/CompliantTextBatchReader.java
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.drill.exec.store.easy.text.compliant.v3;
+package org.apache.drill.exec.store.easy.text.reader;
 
 import java.io.IOException;
 import java.io.InputStream;
@@ -49,7 +49,7 @@ public class CompliantTextBatchReader implements ManagedReader<ColumnsSchemaNego
   private static final int WHITE_SPACE_BUFFER = 64 * 1024;
 
   // settings to be used while parsing
-  private final TextParsingSettingsV3 settings;
+  private final TextParsingSettings settings;
   // Chunk of the file to be read by this reader
   private FileSplit split;
   // text reader implementation
@@ -62,7 +62,7 @@ public class CompliantTextBatchReader implements ManagedReader<ColumnsSchemaNego
 
   private RowSetLoader writer;
 
-  public CompliantTextBatchReader(TextParsingSettingsV3 settings) {
+  public CompliantTextBatchReader(TextParsingSettings settings) {
     this.settings = settings;
 
     // Validate. Otherwise, these problems show up later as a data
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/FieldVarCharOutput.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/FieldVarCharOutput.java
similarity index 97%
rename from exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/FieldVarCharOutput.java
rename to exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/FieldVarCharOutput.java
index e291b3f..048c982 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/FieldVarCharOutput.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/FieldVarCharOutput.java
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.drill.exec.store.easy.text.compliant.v3;
+package org.apache.drill.exec.store.easy.text.reader;
 
 import org.apache.drill.exec.physical.rowSet.RowSetLoader;
 import org.apache.drill.exec.record.metadata.TupleMetadata;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/HeaderBuilder.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/HeaderBuilder.java
similarity index 99%
rename from exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/HeaderBuilder.java
rename to exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/HeaderBuilder.java
index 62eafc8..2fb0ffc 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/HeaderBuilder.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/HeaderBuilder.java
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.drill.exec.store.easy.text.compliant.v3;
+package org.apache.drill.exec.store.easy.text.reader;
 
 import java.nio.BufferOverflowException;
 import java.nio.ByteBuffer;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/RepeatedVarCharOutput.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/RepeatedVarCharOutput.java
similarity index 98%
rename from exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/RepeatedVarCharOutput.java
rename to exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/RepeatedVarCharOutput.java
index 4c1a0b4..b8551e0 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/RepeatedVarCharOutput.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/RepeatedVarCharOutput.java
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.drill.exec.store.easy.text.compliant.v3;
+package org.apache.drill.exec.store.easy.text.reader;
 
 import org.apache.drill.common.exceptions.UserException;
 import org.apache.drill.exec.physical.rowSet.RowSetLoader;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/StreamFinishedPseudoException.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/StreamFinishedPseudoException.java
similarity index 83%
rename from exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/StreamFinishedPseudoException.java
rename to exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/StreamFinishedPseudoException.java
index f2a32b9..0ac2cc6 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/StreamFinishedPseudoException.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/StreamFinishedPseudoException.java
@@ -15,15 +15,14 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.drill.exec.store.easy.text.compliant;
+package org.apache.drill.exec.store.easy.text.reader;
 
+@SuppressWarnings("serial")
 class StreamFinishedPseudoException extends RuntimeException {
-
-  public static final StreamFinishedPseudoException INSTANCE = new StreamFinishedPseudoException();
+  public static final StreamFinishedPseudoException INSTANCE =
+        new StreamFinishedPseudoException();
 
   private StreamFinishedPseudoException() {
     super("", null, false, true);
-
   }
-
 }
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/TextInput.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/TextInput.java
similarity index 97%
rename from exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/TextInput.java
rename to exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/TextInput.java
index 28ddd07..3e05e58 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/TextInput.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/TextInput.java
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.drill.exec.store.easy.text.compliant.v3;
+package org.apache.drill.exec.store.easy.text.reader;
 
 import static org.apache.drill.exec.memory.BoundsChecking.rangeCheck;
 
@@ -41,7 +41,7 @@ final class TextInput {
 
   private final byte[] lineSeparator;
   private final byte normalizedLineSeparator;
-  private final TextParsingSettingsV3 settings;
+  private final TextParsingSettings settings;
 
   private long lineCount;
   private long charCount;
@@ -91,7 +91,7 @@ final class TextInput {
    * {@link Format#getNormalizedNewline()}) that is used to replace any
    * lineSeparator sequence found in the input.
    */
-  public TextInput(TextParsingSettingsV3 settings, InputStream input, DrillBuf readBuffer, long startPos, long endPos) {
+  public TextInput(TextParsingSettings settings, InputStream input, DrillBuf readBuffer, long startPos, long endPos) {
     this.lineSeparator = settings.getNewLineDelimiter();
     byte normalizedLineSeparator = settings.getNormalizedNewLine();
     Preconditions.checkArgument(input instanceof Seekable, "Text input only supports an InputStream that supports Seekable.");
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/TextOutput.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/TextOutput.java
similarity index 97%
rename from exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/TextOutput.java
rename to exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/TextOutput.java
index 48c1849..71d4731 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/TextOutput.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/TextOutput.java
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.drill.exec.store.easy.text.compliant.v3;
+package org.apache.drill.exec.store.easy.text.reader;
 
 
 /**
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/TextParsingContext.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/TextParsingContext.java
similarity index 97%
rename from exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/TextParsingContext.java
rename to exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/TextParsingContext.java
index 86cad4c..3f6dbeb 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/TextParsingContext.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/TextParsingContext.java
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.drill.exec.store.easy.text.compliant.v3;
+package org.apache.drill.exec.store.easy.text.reader;
 
 import java.io.IOException;
 
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/TextParsingSettingsV3.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/TextParsingSettings.java
similarity index 97%
rename from exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/TextParsingSettingsV3.java
rename to exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/TextParsingSettings.java
index 12bbf42..a91a8ab 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/TextParsingSettingsV3.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/TextParsingSettings.java
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.drill.exec.store.easy.text.compliant.v3;
+package org.apache.drill.exec.store.easy.text.reader;
 
 import org.apache.drill.exec.record.metadata.TupleMetadata;
 import org.apache.drill.exec.server.options.OptionManager;
@@ -24,8 +24,7 @@ import org.apache.drill.exec.store.easy.text.TextFormatPlugin;
 import org.apache.drill.exec.store.easy.text.TextFormatPlugin.TextFormatConfig;
 import org.apache.drill.shaded.guava.com.google.common.base.Charsets;
 
-// TODO: Remove the "V3" suffix once the V2 version is retired.
-public class TextParsingSettingsV3 {
+public class TextParsingSettings {
 
   private final String emptyValue = null;
   private final boolean parseUnescapedQuotes = true;
@@ -65,7 +64,7 @@ public class TextParsingSettingsV3 {
    * <tt>`csv`</tt> config has no headers. But, if the user has a ".csv"
    * file with headers, the user can just customize the table properties.
    */
-  public TextParsingSettingsV3(TextFormatConfig config,
+  public TextParsingSettings(TextFormatConfig config,
       EasySubScan scan, OptionManager options) {
     TupleMetadata providedSchema = scan.getSchema();
     boolean extractHeaders = config.isHeaderExtractionEnabled();
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/TextReader.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/TextReader.java
similarity index 95%
rename from exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/TextReader.java
rename to exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/TextReader.java
index 78adda0..0ce856e 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/TextReader.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/TextReader.java
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.drill.exec.store.easy.text.compliant.v3;
+package org.apache.drill.exec.store.easy.text.reader;
 
 import java.io.IOException;
 
@@ -42,7 +42,7 @@ public final class TextReader {
 
   private final TextParsingContext context;
 
-  private final TextParsingSettingsV3 settings;
+  private final TextParsingSettings settings;
 
   private final TextInput input;
   private final TextOutput output;
@@ -70,14 +70,14 @@ public final class TextReader {
   private final byte newLine;
 
   /**
-   * The CsvParser supports all settings provided by {@link TextParsingSettingsV3},
+   * The CsvParser supports all settings provided by {@link TextParsingSettings},
    * and requires this configuration to be properly initialized.
    * @param settings  the parser configuration
    * @param input  input stream
    * @param output  interface to produce output record batch
    * @param workBuf  working buffer to handle whitespace
    */
-  public TextReader(TextParsingSettingsV3 settings, TextInput input, TextOutput output, DrillBuf workBuf) {
+  public TextReader(TextParsingSettings settings, TextInput input, TextOutput output, DrillBuf workBuf) {
     this.context = new TextParsingContext(input, output);
     this.workBuf = workBuf;
     this.settings = settings;
@@ -135,10 +135,8 @@ public final class TextReader {
     output.startRecord();
     int fieldsWritten = 0;
     try {
-      @SuppressWarnings("unused")
-      boolean earlyTerm = false;
       while (ch != newLine) {
-        earlyTerm = ! parseField();
+        parseField();
         fieldsWritten++;
         if (ch != newLine) {
           ch = input.nextChar();
@@ -148,15 +146,6 @@ public final class TextReader {
             break;
           }
         }
-
-        // Disabling early termination. See DRILL-5914
-
-//        if (earlyTerm) {
-//          if (ch != newLine) {
-//            input.skipLines(1);
-//          }
-//          break;
-//        }
       }
       output.finishRecord();
     } catch (StreamFinishedPseudoException e) {
@@ -268,11 +257,9 @@ public final class TextReader {
         if (prev == quoteEscape) {
           output.append(prev);
         }
-        if (prev == quote) { // unescaped quote detected
+        else if (prev == quote) { // unescaped quote detected
           if (parseUnescapedQuotes) {
-            output.append(quote);
-            output.append(ch);
-            parseQuotedValue(ch);
+            output.append(prev);
             break;
           } else {
             throw new TextParsingException(
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/package-info.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/package-info.java
similarity index 93%
rename from exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/package-info.java
rename to exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/package-info.java
index aced5ad..970ef85 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/v3/package-info.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/reader/package-info.java
@@ -19,4 +19,4 @@
  * Version 3 of the text reader. Hosts the "compliant" text reader on
  * the row set framework.
  */
-package org.apache.drill.exec.store.easy.text.compliant.v3;
+package org.apache.drill.exec.store.easy.text.reader;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/text/DrillTextRecordWriter.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/writer/TextRecordWriter.java
similarity index 94%
rename from exec/java-exec/src/main/java/org/apache/drill/exec/store/text/DrillTextRecordWriter.java
rename to exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/writer/TextRecordWriter.java
index 83a00bd..a114fd4 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/text/DrillTextRecordWriter.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/writer/TextRecordWriter.java
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.drill.exec.store.text;
+package org.apache.drill.exec.store.easy.text.writer;
 
 import java.io.DataOutputStream;
 import java.io.IOException;
@@ -24,18 +24,17 @@ import java.util.List;
 import java.util.Map;
 
 import org.apache.drill.exec.memory.BufferAllocator;
-import org.apache.drill.exec.store.StorageStrategy;
 import org.apache.drill.exec.store.EventBasedRecordWriter.FieldConverter;
+import org.apache.drill.exec.store.StorageStrategy;
 import org.apache.drill.exec.store.StringOutputRecordWriter;
 import org.apache.drill.exec.vector.complex.reader.FieldReader;
+import org.apache.drill.shaded.guava.com.google.common.base.Joiner;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
 
-import org.apache.drill.shaded.guava.com.google.common.base.Joiner;
-
-public class DrillTextRecordWriter extends StringOutputRecordWriter {
-  static final org.slf4j.Logger logger = org.slf4j.LoggerFactory.getLogger(DrillTextRecordWriter.class);
+public class TextRecordWriter extends StringOutputRecordWriter {
+  static final org.slf4j.Logger logger = org.slf4j.LoggerFactory.getLogger(TextRecordWriter.class);
 
   private final StorageStrategy storageStrategy;
 
@@ -47,7 +46,6 @@ public class DrillTextRecordWriter extends StringOutputRecordWriter {
   private String fieldDelimiter;
   private String extension;
 
-  private static String eol = System.getProperty("line.separator");
   private int index;
   private PrintStream stream = null;
   private FileSystem fs = null;
@@ -58,7 +56,7 @@ public class DrillTextRecordWriter extends StringOutputRecordWriter {
 
   private Configuration fsConf;
 
-  public DrillTextRecordWriter(BufferAllocator allocator, StorageStrategy storageStrategy, Configuration fsConf) {
+  public TextRecordWriter(BufferAllocator allocator, StorageStrategy storageStrategy, Configuration fsConf) {
     super(allocator);
     this.storageStrategy = storageStrategy == null ? StorageStrategy.DEFAULT : storageStrategy;
     this.fsConf = new Configuration(fsConf);
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/text/DrillTextRecordReader.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/text/DrillTextRecordReader.java
deleted file mode 100644
index 0db17fb..0000000
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/text/DrillTextRecordReader.java
+++ /dev/null
@@ -1,244 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.drill.exec.store.text;
-
-import java.io.IOException;
-import java.util.Collections;
-import java.util.List;
-
-import javax.annotation.Nullable;
-
-import org.apache.drill.shaded.guava.com.google.common.base.Predicate;
-import org.apache.drill.shaded.guava.com.google.common.collect.Iterables;
-
-import org.apache.drill.common.exceptions.DrillRuntimeException;
-import org.apache.drill.common.exceptions.ExecutionSetupException;
-import org.apache.drill.common.expression.FieldReference;
-import org.apache.drill.common.expression.SchemaPath;
-import org.apache.drill.common.types.TypeProtos;
-import org.apache.drill.common.types.Types;
-import org.apache.drill.exec.ExecConstants;
-import org.apache.drill.exec.ops.FragmentContext;
-import org.apache.drill.exec.ops.OperatorContext;
-import org.apache.drill.exec.physical.impl.OutputMutator;
-import org.apache.drill.exec.record.MaterializedField;
-import org.apache.drill.exec.store.AbstractRecordReader;
-import org.apache.drill.exec.vector.RepeatedVarCharVector;
-import org.apache.drill.exec.vector.ValueVector;
-import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.io.LongWritable;
-import org.apache.hadoop.io.Text;
-import org.apache.hadoop.mapred.FileSplit;
-import org.apache.hadoop.mapred.JobConf;
-import org.apache.hadoop.mapred.Reporter;
-import org.apache.hadoop.mapred.TextInputFormat;
-
-import org.apache.drill.shaded.guava.com.google.common.base.Preconditions;
-import org.apache.drill.shaded.guava.com.google.common.collect.Lists;
-
-public class DrillTextRecordReader extends AbstractRecordReader {
-  private static final org.slf4j.Logger logger = org.slf4j.LoggerFactory.getLogger(DrillTextRecordReader.class);
-
-  private static final String COL_NAME = "columns";
-
-  private org.apache.hadoop.mapred.RecordReader<LongWritable, Text> reader;
-  private final List<ValueVector> vectors = Lists.newArrayList();
-  private byte delimiter;
-  private FieldReference ref = new FieldReference(COL_NAME);
-  private RepeatedVarCharVector vector;
-  private List<Integer> columnIds = Lists.newArrayList();
-  private LongWritable key;
-  private Text value;
-  private int numCols = 0;
-  private FileSplit split;
-  private long totalRecordsRead;
-
-  public DrillTextRecordReader(FileSplit split, Configuration fsConf, FragmentContext context,
-      char delimiter, List<SchemaPath> columns) {
-    this.delimiter = (byte) delimiter;
-    this.split = split;
-    setColumns(columns);
-
-    if (!isStarQuery()) {
-      String pathStr;
-      for (SchemaPath path : columns) {
-        assert path.getRootSegment().isNamed();
-        pathStr = path.getRootSegment().getPath();
-        Preconditions.checkArgument(COL_NAME.equals(pathStr) || (SchemaPath.DYNAMIC_STAR.equals(pathStr) && path.getRootSegment().getChild() == null),
-            "Selected column(s) must have name 'columns' or must be plain '*'");
-
-        if (path.getRootSegment().getChild() != null) {
-          Preconditions.checkArgument(path.getRootSegment().getChild().isArray(), "Selected column must be an array index");
-          int index = path.getRootSegment().getChild().getArraySegment().getIndex();
-          columnIds.add(index);
-        }
-      }
-      Collections.sort(columnIds);
-      numCols = columnIds.size();
-    }
-
-    TextInputFormat inputFormat = new TextInputFormat();
-    JobConf job = new JobConf(fsConf);
-    job.setInt("io.file.buffer.size", context.getConfig().getInt(ExecConstants.TEXT_LINE_READER_BUFFER_SIZE));
-    job.setInputFormat(inputFormat.getClass());
-    try {
-      reader = inputFormat.getRecordReader(split, job, Reporter.NULL);
-      key = reader.createKey();
-      value = reader.createValue();
-      totalRecordsRead = 0;
-    } catch (Exception e) {
-      handleAndRaise("Failure in creating record reader", e);
-    }
-  }
-
-  @Override
-  protected List<SchemaPath> getDefaultColumnsToRead() {
-    return DEFAULT_TEXT_COLS_TO_READ;
-  }
-
-  @Override
-  public boolean isStarQuery() {
-    return super.isStarQuery() || Iterables.tryFind(getColumns(), new Predicate<SchemaPath>() {
-      private final SchemaPath COLUMNS = SchemaPath.getSimplePath("columns");
-      @Override
-      public boolean apply(@Nullable SchemaPath path) {
-        return path.equals(COLUMNS);
-      }
-    }).isPresent();
-  }
-
-  @Override
-  public void setup(OperatorContext context, OutputMutator output) throws ExecutionSetupException {
-    MaterializedField field = MaterializedField.create(ref.getAsNamePart().getName(), Types.repeated(TypeProtos.MinorType.VARCHAR));
-    try {
-      vector = output.addField(field, RepeatedVarCharVector.class);
-    } catch (Exception e) {
-      handleAndRaise("Failure in setting up reader", e);
-    }
-  }
-
-  protected void handleAndRaise(String s, Exception e) {
-    String message = "Error in text record reader.\nMessage: " + s +
-      "\nSplit information:\n\tPath: " + split.getPath() +
-      "\n\tStart: " + split.getStart() +
-      "\n\tLength: " + split.getLength();
-    throw new DrillRuntimeException(message, e);
-  }
-
-  @Override
-  public int next() {
-//    logger.debug("vector value capacity {}", vector.getValueCapacity());
-//    logger.debug("vector byte capacity {}", vector.getByteCapacity());
-    int batchSize = 0;
-    try {
-      int recordCount = 0;
-      final RepeatedVarCharVector.Mutator mutator = vector.getMutator();
-      while (recordCount < Character.MAX_VALUE && batchSize < 200*1000 && reader.next(key, value)) {
-        int start;
-        int end = -1;
-
-        // index of the scanned field
-        int p = 0;
-        int i = 0;
-        mutator.startNewValue(recordCount);
-        // Process each field in this line
-        while (end < value.getLength() - 1) {
-          if(numCols > 0 && p >= numCols) {
-            break;
-          }
-          start = end;
-          if (delimiter == '\n') {
-            end = value.getLength();
-          } else {
-            end = find(value, delimiter, start + 1);
-            if (end == -1) {
-              end = value.getLength();
-            }
-          }
-          if (numCols > 0 && i++ < columnIds.get(p)) {
-            mutator.addSafe(recordCount, value.getBytes(), start + 1, 0);
-            continue;
-          }
-          p++;
-          mutator.addSafe(recordCount, value.getBytes(), start + 1, end - start - 1);
-          batchSize += end - start;
-        }
-        recordCount++;
-        totalRecordsRead++;
-      }
-      for (final ValueVector v : vectors) {
-        v.getMutator().setValueCount(recordCount);
-      }
-      mutator.setValueCount(recordCount);
-      // logger.debug("text scan batch size {}", batchSize);
-      return recordCount;
-    } catch(Exception e) {
-      close();
-      handleAndRaise("Failure while parsing text. Parser was at record: " + (totalRecordsRead + 1), e);
-    }
-
-    // this is never reached
-    return 0;
-  }
-
-  /**
-   * Returns the index within the text of the first occurrence of delimiter, starting the search at the specified index.
-   *
-   * @param  text  the text being searched
-   * @param  delimiter the delimiter
-   * @param  start the index to start searching
-   * @return      the first occurrence of delimiter, starting the search at the specified index
-   */
-  public int find(Text text, byte delimiter, int start) {
-    int len = text.getLength();
-    int p = start;
-    byte[] bytes = text.getBytes();
-    boolean inQuotes = false;
-    while (p < len) {
-      if ('\"' == bytes[p]) {
-        inQuotes = !inQuotes;
-      }
-      if (!inQuotes && bytes[p] == delimiter) {
-        return p;
-      }
-      p++;
-    }
-    return -1;
-  }
-
-  @Override
-  public void close() {
-    try {
-      if (reader != null) {
-        reader.close();
-        reader = null;
-      }
-    } catch (IOException e) {
-      logger.warn("Exception closing reader: {}", e);
-    }
-  }
-
-  @Override
-  public String toString() {
-    return "DrillTextRecordReader[File=" + split.getPath()
-        + ", Record=" + (totalRecordsRead + 1)
-        + ", Start=" + split.getStart()
-        + ", Length=" + split.getLength()
-        + "]";
-  }
-}
diff --git a/exec/java-exec/src/test/java/org/apache/drill/TestSelectWithOption.java b/exec/java-exec/src/test/java/org/apache/drill/TestSelectWithOption.java
index 7bc65fc..c5c283d 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/TestSelectWithOption.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/TestSelectWithOption.java
@@ -170,7 +170,7 @@ public class TestSelectWithOption extends BaseTestQuery {
     // It seems that a parameter can not be called "escape"
     testWithResult(format("select columns from table(%s(`escape` => '$', type => 'TeXT', fieldDelimiter => '|', quote => '@'))", quoteTableName),
         listOf("b", "0"),
-        listOf("b$@c", "1")); // shouldn't $ be removed here?
+        listOf("b@c", "1"));
   }
 
   @Test
@@ -230,9 +230,14 @@ public class TestSelectWithOption extends BaseTestQuery {
     String jsonTableName = genCSVTable("testVariationsJSON",
         "{\"columns\": [\"f\",\"g\"]}");
     // the extension is actually csv
-    testWithResult(format("select columns from %s", jsonTableName),
-        listOf("{\"columns\": [\"f\"", "g\"]}\n")
-        );
+    // Don't try to read the CSV file, however, as it does not
+    // contain proper quotes for CSV.
+    // File contents:
+    // {"columns": ["f","g"]}
+    // CSV would require:
+    // "{""columns"": [""f"",""g""]}"
+    // A bug in older versions appeared to have the perverse
+    // effect of allowing the above to kinds-sorta work.
     String[] jsonQueries = {
         format("select columns from table(%s(type => 'JSON'))", jsonTableName),
         // we can use named format plugin configurations too!
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/store/TestImplicitFileColumns.java b/exec/java-exec/src/test/java/org/apache/drill/exec/store/TestImplicitFileColumns.java
index 07ffba0..403aab1 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/store/TestImplicitFileColumns.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/store/TestImplicitFileColumns.java
@@ -17,21 +17,21 @@
  */
 package org.apache.drill.exec.store;
 
-import org.apache.drill.shaded.guava.com.google.common.base.Charsets;
-import org.apache.drill.shaded.guava.com.google.common.io.Files;
-import org.apache.drill.test.BaseTestQuery;
+import java.io.File;
+import java.nio.file.Path;
+import java.nio.file.Paths;
+
 import org.apache.drill.common.types.TypeProtos;
 import org.apache.drill.exec.record.BatchSchema;
 import org.apache.drill.exec.record.metadata.SchemaBuilder;
 import org.apache.drill.exec.util.JsonStringArrayList;
 import org.apache.drill.exec.util.Text;
+import org.apache.drill.shaded.guava.com.google.common.base.Charsets;
+import org.apache.drill.shaded.guava.com.google.common.io.Files;
+import org.apache.drill.test.BaseTestQuery;
 import org.junit.BeforeClass;
 import org.junit.Test;
 
-import java.io.File;
-import java.nio.file.Path;
-import java.nio.file.Paths;
-
 public class TestImplicitFileColumns extends BaseTestQuery {
   public static final String CSV = "csv";
   public static final String MAIN = "main";
@@ -199,9 +199,9 @@ public class TestImplicitFileColumns extends BaseTestQuery {
   @Test
   public void testStarColumnCsv() throws Exception {
     final BatchSchema expectedSchema = new SchemaBuilder()
+        .addArray("columns", TypeProtos.MinorType.VARCHAR)
         .addNullable("dir0", TypeProtos.MinorType.VARCHAR)
         .addNullable("dir1", TypeProtos.MinorType.VARCHAR)
-        .addArray("columns", TypeProtos.MinorType.VARCHAR)
         .build();
 
     testBuilder()
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/BaseCsvTest.java b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/BaseCsvTest.java
index 3aa5584..2071546 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/BaseCsvTest.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/BaseCsvTest.java
@@ -80,14 +80,6 @@ public class BaseCsvTest extends ClusterTest {
     buildFile(new File(nestedDir, NESTED_FILE), secondFile);
   }
 
-  protected void enableV3(boolean enable) {
-    client.alterSession(ExecConstants.ENABLE_V3_TEXT_READER_KEY, enable);
-  }
-
-  protected void resetV3() {
-    client.resetSession(ExecConstants.ENABLE_V3_TEXT_READER_KEY);
-  }
-
   protected void enableMultiScan() {
 
     // Special test-only feature to force even small scans
@@ -153,12 +145,10 @@ public class BaseCsvTest extends ClusterTest {
   }
 
   protected void enableSchemaSupport() {
-    enableV3(true);
     enableSchema(true);
   }
 
   protected void resetSchemaSupport() {
-    resetV3();
     resetSchema();
   }
 }
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/store/text/TestCsvHeader.java b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvHeader.java
similarity index 99%
rename from exec/java-exec/src/test/java/org/apache/drill/exec/store/text/TestCsvHeader.java
rename to exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvHeader.java
index fdef252..d190a63 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/store/text/TestCsvHeader.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvHeader.java
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.drill.exec.store.text;
+package org.apache.drill.exec.store.easy.text.compliant;
 
 import org.apache.drill.shaded.guava.com.google.common.collect.Lists;
 import org.apache.drill.test.BaseTestQuery;
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvIgnoreHeaders.java b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvIgnoreHeaders.java
index 5a52664..dea2c9e 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvIgnoreHeaders.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvIgnoreHeaders.java
@@ -33,9 +33,8 @@ import org.junit.BeforeClass;
 import org.junit.Test;
 import org.junit.experimental.categories.Category;
 
-// CSV reader now hosted on the row set framework
 @Category(RowSetTests.class)
-public class TestCsvIgnoreHeaders  extends BaseCsvTest{
+public class TestCsvIgnoreHeaders extends BaseCsvTest{
 
   private static String withHeaders[] = {
       "a,b,c",
@@ -59,17 +58,6 @@ public class TestCsvIgnoreHeaders  extends BaseCsvTest{
   public void testColumns() throws IOException {
     String fileName = "simple.csv";
     buildFile(fileName, withHeaders);
-    try {
-      enableV3(false);
-      doTestColumns(fileName);
-      enableV3(true);
-      doTestColumns(fileName);
-    } finally {
-      resetV3();
-    }
-  }
-
-  private void doTestColumns(String fileName) throws IOException {
     String sql = "SELECT columns FROM `dfs.data`.`%s`";
     RowSet actual = client.queryBuilder().sql(sql, fileName).rowSet();
 
@@ -86,17 +74,6 @@ public class TestCsvIgnoreHeaders  extends BaseCsvTest{
 
   @Test
   public void testRaggedRows() throws IOException {
-    try {
-      enableV3(false);
-      doTestRaggedRows();
-      enableV3(true);
-      doTestRaggedRows();
-    } finally {
-      resetV3();
-    }
-  }
-
-  private void doTestRaggedRows() throws IOException {
     String fileName = "ragged.csv";
     TestCsvWithHeaders.buildFile(new File(testDir, fileName), raggedRows);
     String sql = "SELECT columns FROM `dfs.data`.`%s`";
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvTableProperties.java b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvTableProperties.java
index a540694..7c40128 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvTableProperties.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvTableProperties.java
@@ -23,6 +23,7 @@ import java.io.File;
 import java.io.FileWriter;
 import java.io.PrintWriter;
 
+import org.apache.drill.TestSelectWithOption;
 import org.apache.drill.categories.RowSetTests;
 import org.apache.drill.common.types.TypeProtos.MinorType;
 import org.apache.drill.exec.record.metadata.SchemaBuilder;
@@ -47,6 +48,9 @@ import org.junit.experimental.categories.Category;
  * is provided, the text format plugin will create columns
  * using that schema rather than using the "columns" array
  * column.
+ *
+ * @see {@link TestSelectWithOption} for similar tests using table
+ * properties within SQL
  */
 
 @Category(RowSetTests.class)
@@ -448,4 +452,67 @@ public class TestCsvTableProperties extends BaseCsvTest {
       resetSchemaSupport();
     }
   }
+
+  private static String messyQuotesData[] = {
+      "first\"field\"here,another \"field",
+      "end quote\",another\"",
+      "many\"\"\"\",more\"\"",
+      "\"not\"end\",\"\"wtf\" \"",
+      "\"newline\nhere\",\"and here\"\"\n\""
+    };
+
+  /**
+   * The legacy "V2" text reader had special handling for quotes
+   * that appear inside fields. Example:<pre><tt>
+   * first"field"here,another "field</tt></pre>
+   * <p>
+   * Since behavior in this case is ill-defined, the reader
+   * apparently treated quotes as normal characters unless the
+   * field started with a quote. There is an option in the UniVocity
+   * code to set this behavior, but it is not exposed in Drill.
+   * So, this test verifies the non-customizable messy quote handling
+   * logic.
+   * <p>
+   * If a field starts with a quote, quoting rules kick in, including
+   * the quote escape, which is, by default, itself a quote. So
+   * <br><code>"foo""bar"</code><br>
+   * is read as
+   * <br><code>foo"bar</code><br>
+   * But, for fields not starting with a quote, the quote escape
+   * is ignored, so:
+   * <br><code>foo""bar</code><br>
+   * is read as
+   * <br><code>foo""bar</code><br>
+   * This seems more like a bug than a feature, but it does appear to be
+   * how the "new" text reader always worked, so the behavior is preserved.
+   * <p>
+   * Also, seems that the text reader supported embedded newlines, even
+   * though such behavior <i><b>will not work</b></i> if the embedded
+   * newline occurs near a split. In this case, the reader will scan
+   * forward to find a record delimiter (a newline by default), will
+   * find the embedded newline, and will read a partial first record.
+   * Again, this appears to be legacy behavior, and so is preserved,
+   * even if broken.
+   * <p>
+   * The key thing is that if the CSV is well-formed (no messy quotes,
+   * properly quoted fields with proper escapes, no embedded newlines)
+   * then things will work OK.
+   */
+
+  @Test
+  public void testMessyQuotes() throws Exception {
+   String tablePath = buildTable("messyQuotes", messyQuotesData);
+    RowSet actual = client.queryBuilder().sql(SELECT_ALL, tablePath).rowSet();
+    TupleMetadata expectedSchema = new SchemaBuilder()
+        .addArray("columns", MinorType.VARCHAR)
+        .buildSchema();
+    RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
+        .addSingleCol(strArray("first\"field\"here", "another \"field"))
+        .addSingleCol(strArray("end quote\"", "another\""))
+        .addSingleCol(strArray("many\"\"\"\"", "more\"\""))
+        .addSingleCol(strArray("not\"end", "\"wtf\" "))
+        .addSingleCol(strArray("newline\nhere", "and here\"\n"))
+        .build();
+    RowSetUtilities.verify(expected, actual);
+  }
 }
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvWithHeaders.java b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvWithHeaders.java
index 7abbf3d..0b0c181 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvWithHeaders.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvWithHeaders.java
@@ -41,47 +41,27 @@ import org.junit.Test;
 import org.junit.experimental.categories.Category;
 
 /**
- * Sanity test of CSV files with headers. Tests both the original
- * "compliant" version and the V3 version based on the row set
- * framework.
+ * Sanity test of CSV files with headers.
  * <p>
- * The CSV reader is a "canary in the coal mine" for many scan features.
- * It turns out that there are several bugs in "V2" (AKA "new text reader")
- * that are fixed in "V3" (the one based on the row set framework), and one
- * that is not yet fixed.
+ * Open issues:
  *
  * <ul>
- * <li>Ragged rows will crash the V2 text reader when headers are used.
- * No V2 test exists as a result. Fixed in V3.</li>
- * <li>DRILL-7083: in V2, if files are nested to 2 levels, but we ask
- * for dir2 (the non-existent third level), the type of dir2 will be
- * nullable INT. In V3, the type is Nullable VARCHAR (just like for the
- * existing partition levels.)</li>
  * <li>DRILL-7080: A query like SELECT *, dir0 produces the result schema
  * of (dir0, a, b, ...) in V2 and (a, b, ... dir0, dir00) in V3. This
  * seems to be a bug in the Project operator.</li>
  * </ul>
  *
- * The V3 tests all demonstrate that the row set scan framework
+ * The tests all demonstrate that the row set scan framework
  * delivers a first empty batch from each scan. I (Paul) had understood
  * that we had an "fast schema" path as the result of the "empty batch"
- * project. However, the V2 reader does not provide the schema-only
+ * project. However, the V2 reader did not provide the schema-only
  * first batch. So, not sure if doing so is a feature, or a bug because
  * things changed. Easy enough to change if we choose to. If so, the
  * tests here would remove the test for that schema-only batch.
- * <p>
- * Tests are run for both V2 and V3. When the results are the same,
- * the test occurs once, wrapped in a "driver" to select V2 or V3 mode.
- * When behavior differs, there are separate tests for V2 and V3.
- * <p>
- * The V2 tests are temporary. Once we accept that V3 is stable, we
- * can remove V2 (and the "old text reader.") The behavior in V3 is
- * more correct, no reason to keep the old, broken behavior.
  *
  * @see {@link TestHeaderBuilder}
  */
 
-// CSV reader now hosted on the row set framework
 @Category(RowSetTests.class)
 public class TestCsvWithHeaders extends BaseCsvTest {
 
@@ -125,17 +105,6 @@ public class TestCsvWithHeaders extends BaseCsvTest {
   @Test
   public void testEmptyFile() throws IOException {
     buildFile(EMPTY_FILE, new String[] {});
-    try {
-      enableV3(false);
-      doTestEmptyFile();
-      enableV3(true);
-      doTestEmptyFile();
-    } finally {
-      resetV3();
-    }
-  }
-
-  private void doTestEmptyFile() throws IOException {
     RowSet rowSet = client.queryBuilder().sql(makeStatement(EMPTY_FILE)).rowSet();
     assertNull(rowSet);
   }
@@ -150,17 +119,6 @@ public class TestCsvWithHeaders extends BaseCsvTest {
   public void testEmptyCsvHeaders() throws IOException {
     buildFile(EMPTY_HEADERS_FILE, emptyHeaders);
     try {
-      enableV3(false);
-      doTestEmptyCsvHeaders();
-      enableV3(true);
-      doTestEmptyCsvHeaders();
-    } finally {
-      resetV3();
-    }
-  }
-
-  private void doTestEmptyCsvHeaders() throws IOException {
-    try {
       client.queryBuilder().sql(makeStatement(EMPTY_HEADERS_FILE)).run();
       fail();
     } catch (Exception e) {
@@ -170,17 +128,6 @@ public class TestCsvWithHeaders extends BaseCsvTest {
 
   @Test
   public void testValidCsvHeaders() throws IOException {
-    try {
-      enableV3(false);
-      doTestValidCsvHeaders();
-      enableV3(true);
-      doTestValidCsvHeaders();
-    } finally {
-      resetV3();
-    }
-  }
-
-  private void doTestValidCsvHeaders() throws IOException {
     RowSet actual = client.queryBuilder().sql(makeStatement(TEST_FILE_NAME)).rowSet();
 
     TupleMetadata expectedSchema = new SchemaBuilder()
@@ -196,17 +143,6 @@ public class TestCsvWithHeaders extends BaseCsvTest {
 
   @Test
   public void testInvalidCsvHeaders() throws IOException {
-    try {
-      enableV3(false);
-      doTestInvalidCsvHeaders();
-      enableV3(true);
-      doTestInvalidCsvHeaders();
-    } finally {
-      resetV3();
-    }
-  }
-
-  private void doTestInvalidCsvHeaders() throws IOException {
     String fileName = "case3.csv";
     buildFile(fileName, invalidHeaders);
     RowSet actual = client.queryBuilder().sql(makeStatement(fileName)).rowSet();
@@ -225,20 +161,9 @@ public class TestCsvWithHeaders extends BaseCsvTest {
     RowSetUtilities.verify(expected, actual);
   }
 
+  // Test fix for DRILL-5590
   @Test
   public void testCsvHeadersCaseInsensitive() throws IOException {
-    try {
-      enableV3(false);
-      doTestCsvHeadersCaseInsensitive();
-      enableV3(true);
-      doTestCsvHeadersCaseInsensitive();
-    } finally {
-      resetV3();
-    }
-  }
-
-  // Test fix for DRILL-5590
-  private void doTestCsvHeadersCaseInsensitive() throws IOException {
     String sql = "SELECT A, b, C FROM `dfs.data`.`%s`";
     RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
 
@@ -258,23 +183,12 @@ public class TestCsvWithHeaders extends BaseCsvTest {
     return "SELECT * FROM `dfs.data`.`" + fileName + "`";
   }
 
-  @Test
-  public void testWildcard() throws IOException {
-    try {
-      enableV3(false);
-      doTestWildcard();
-      enableV3(true);
-      doTestWildcard();
-    } finally {
-      resetV3();
-    }
-  }
-
   /**
    * Verify that the wildcard expands columns to the header names, including
    * case
    */
-  private void doTestWildcard() throws IOException {
+  @Test
+  public void testWildcard() throws IOException {
     String sql = "SELECT * FROM `dfs.data`.`%s`";
     RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
 
@@ -292,85 +206,24 @@ public class TestCsvWithHeaders extends BaseCsvTest {
 
   /**
    * Verify that implicit columns are recognized and populated. Sanity test
-   * of just one implicit column. V2 uses nullable VARCHAR for file
-   * metadata columns.
-   */
-
-  @Test
-  public void testImplicitColsExplicitSelectV2() throws IOException {
-    try {
-      enableV3(false);
-      String sql = "SELECT A, filename FROM `dfs.data`.`%s`";
-      RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
-
-      TupleMetadata expectedSchema = new SchemaBuilder()
-          .add("A", MinorType.VARCHAR)
-          .addNullable("filename", MinorType.VARCHAR)
-          .buildSchema();
-
-      RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-          .addRow("10", TEST_FILE_NAME)
-          .build();
-      RowSetUtilities.verify(expected, actual);
-    } finally {
-      resetV3();
-    }
-  }
-
-  /**
-   * Verify that implicit columns are recognized and populated. Sanity test
    * of just one implicit column. V3 uses non-nullable VARCHAR for file
    * metadata columns.
    */
 
   @Test
-  public void testImplicitColsExplicitSelectV3() throws IOException {
-    try {
-      enableV3(true);
-      String sql = "SELECT A, filename FROM `dfs.data`.`%s`";
-      RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
-
-      TupleMetadata expectedSchema = new SchemaBuilder()
-          .add("A", MinorType.VARCHAR)
-          .add("filename", MinorType.VARCHAR)
-          .buildSchema();
-
-      RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-          .addRow("10", TEST_FILE_NAME)
-          .build();
-      RowSetUtilities.verify(expected, actual);
-    } finally {
-      resetV3();
-    }
-  }
+  public void testImplicitColsExplicitSelect() throws IOException {
+    String sql = "SELECT A, filename FROM `dfs.data`.`%s`";
+    RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
 
-  /**
-   * Verify that implicit columns are recognized and populated. Sanity test
-   * of just one implicit column. V2 uses nullable VARCHAR for file
-   * metadata columns.
-   */
+    TupleMetadata expectedSchema = new SchemaBuilder()
+        .add("A", MinorType.VARCHAR)
+        .add("filename", MinorType.VARCHAR)
+        .buildSchema();
 
-  @Test
-  public void testImplicitColWildcardV2() throws IOException {
-    try {
-      enableV3(false);
-      String sql = "SELECT *, filename FROM `dfs.data`.`%s`";
-      RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
-
-      TupleMetadata expectedSchema = new SchemaBuilder()
-          .add("a", MinorType.VARCHAR)
-          .add("b", MinorType.VARCHAR)
-          .add("c", MinorType.VARCHAR)
-          .addNullable("filename", MinorType.VARCHAR)
-          .buildSchema();
-
-      RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-          .addRow("10", "foo", "bar", TEST_FILE_NAME)
-          .build();
-      RowSetUtilities.verify(expected, actual);
-    } finally {
-      resetV3();
-    }
+    RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
+        .addRow("10", TEST_FILE_NAME)
+        .build();
+    RowSetUtilities.verify(expected, actual);
   }
 
   /**
@@ -380,41 +233,25 @@ public class TestCsvWithHeaders extends BaseCsvTest {
    */
 
   @Test
-  public void testImplicitColWildcardV3() throws IOException {
-    try {
-      enableV3(true);
-      String sql = "SELECT *, filename FROM `dfs.data`.`%s`";
-      RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
-
-      TupleMetadata expectedSchema = new SchemaBuilder()
-          .add("a", MinorType.VARCHAR)
-          .add("b", MinorType.VARCHAR)
-          .add("c", MinorType.VARCHAR)
-          .add("filename", MinorType.VARCHAR)
-          .buildSchema();
-
-      RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-          .addRow("10", "foo", "bar", TEST_FILE_NAME)
-          .build();
-      RowSetUtilities.verify(expected, actual);
-    } finally {
-      resetV3();
-    }
+  public void testImplicitColWildcard() throws IOException {
+    String sql = "SELECT *, filename FROM `dfs.data`.`%s`";
+    RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
+
+    TupleMetadata expectedSchema = new SchemaBuilder()
+        .add("a", MinorType.VARCHAR)
+        .add("b", MinorType.VARCHAR)
+        .add("c", MinorType.VARCHAR)
+        .add("filename", MinorType.VARCHAR)
+        .buildSchema();
+
+    RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
+        .addRow("10", "foo", "bar", TEST_FILE_NAME)
+        .build();
+    RowSetUtilities.verify(expected, actual);
   }
 
   @Test
   public void testColsWithWildcard() throws IOException {
-    try {
-      enableV3(false);
-      doTestColsWithWildcard();
-      enableV3(true);
-      doTestColsWithWildcard();
-    } finally {
-      resetV3();
-    }
-  }
-
-  private void doTestColsWithWildcard() throws IOException {
     String sql = "SELECT *, a as d FROM `dfs.data`.`%s`";
     RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
 
@@ -432,73 +269,30 @@ public class TestCsvWithHeaders extends BaseCsvTest {
   }
 
   /**
-   * V2 does not allow explicit use of dir0, dir1, etc. columns for a non-partitioned
-   * file. Treated as undefined nullable int columns.
-   */
-
-  @Test
-  public void testPartitionColsExplicitV2() throws IOException {
-    try {
-      enableV3(false);
-      String sql = "SELECT a, dir0, dir5 FROM `dfs.data`.`%s`";
-      RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
-
-      TupleMetadata expectedSchema = new SchemaBuilder()
-          .add("a", MinorType.VARCHAR)
-          .addNullable("dir0", MinorType.INT)
-          .addNullable("dir5", MinorType.INT)
-          .buildSchema();
-
-      RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-          .addRow("10", null, null)
-          .build();
-      RowSetUtilities.verify(expected, actual);
-    } finally {
-      resetV3();
-    }
-  }
-
-  /**
    * V3 allows the use of partition columns, even for a non-partitioned file.
    * The columns are null of type Nullable VARCHAR. This is area of Drill
    * is a bit murky: it seems reasonable to support partition columns consistently
    * rather than conditionally based on the structure of the input.
    */
   @Test
-  public void testPartitionColsExplicitV3() throws IOException {
-    try {
-      enableV3(true);
-      String sql = "SELECT a, dir0, dir5 FROM `dfs.data`.`%s`";
-      RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
-
-      TupleMetadata expectedSchema = new SchemaBuilder()
-          .add("a", MinorType.VARCHAR)
-          .addNullable("dir0", MinorType.VARCHAR)
-          .addNullable("dir5", MinorType.VARCHAR)
-          .buildSchema();
-
-      RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-          .addRow("10", null, null)
-          .build();
-      RowSetUtilities.verify(expected, actual);
-    } finally {
-      resetV3();
-    }
+  public void testPartitionColsExplicit() throws IOException {
+    String sql = "SELECT a, dir0, dir5 FROM `dfs.data`.`%s`";
+    RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
+
+    TupleMetadata expectedSchema = new SchemaBuilder()
+        .add("a", MinorType.VARCHAR)
+        .addNullable("dir0", MinorType.VARCHAR)
+        .addNullable("dir5", MinorType.VARCHAR)
+        .buildSchema();
+
+    RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
+        .addRow("10", null, null)
+        .build();
+    RowSetUtilities.verify(expected, actual);
   }
 
   @Test
   public void testDupColumn() throws IOException {
-    try {
-      enableV3(false);
-      doTestDupColumn();
-      enableV3(true);
-      doTestDupColumn();
-    } finally {
-      resetV3();
-    }
-  }
-
-  private void doTestDupColumn() throws IOException {
     String sql = "SELECT a, b, a FROM `dfs.data`.`%s`";
     RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
 
@@ -514,91 +308,31 @@ public class TestCsvWithHeaders extends BaseCsvTest {
     RowSetUtilities.verify(expected, actual);
   }
 
-  // This test cannot be run for V2. The data gets corrupted and we get
-  // internal errors.
-
   /**
    * Test that ragged rows result in the "missing" columns being filled
    * in with the moral equivalent of a null column for CSV: a blank string.
    */
   @Test
-  public void testRaggedRowsV3() throws IOException {
-    try {
-      enableV3(true);
-      String fileName = "case4.csv";
-      buildFile(fileName, raggedRows);
-      RowSet actual = client.queryBuilder().sql(makeStatement(fileName)).rowSet();
-
-      TupleMetadata expectedSchema = new SchemaBuilder()
-          .add("a", MinorType.VARCHAR)
-          .add("b", MinorType.VARCHAR)
-          .add("c", MinorType.VARCHAR)
-          .buildSchema();
-      RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-          .addRow("10", "dino", "")
-          .addRow("20", "foo", "bar")
-          .addRow("30", "", "")
-          .build();
-      RowSetUtilities.verify(expected, actual);
-    } finally {
-      resetV3();
-    }
-  }
+  public void testRaggedRows() throws IOException {
+    String fileName = "case4.csv";
+    buildFile(fileName, raggedRows);
+    RowSet actual = client.queryBuilder().sql(makeStatement(fileName)).rowSet();
 
-  /**
-   * Test partition expansion. Because the two files are read in the
-   * same scan operator, the schema is consistent. See
-   * {@link TestPartitionRace} for the multi-threaded race where all
-   * hell breaks loose.
-   * <p>
-   * V2, since Drill 1.12, puts partition columns ahead of data columns.
-   */
-  @Test
-  public void testPartitionExpansionV2() throws IOException {
-    try {
-      enableV3(false);
-
-      String sql = "SELECT * FROM `dfs.data`.`%s`";
-      Iterator<DirectRowSet> iter = client.queryBuilder().sql(sql, PART_DIR).rowSetIterator();
-
-      TupleMetadata expectedSchema = new SchemaBuilder()
-          .addNullable("dir0", MinorType.VARCHAR)
-          .add("a", MinorType.VARCHAR)
-          .add("b", MinorType.VARCHAR)
-          .add("c", MinorType.VARCHAR)
-          .buildSchema();
-
-      // Read the two batches.
-
-      for (int i = 0; i < 2; i++) {
-        assertTrue(iter.hasNext());
-        RowSet rowSet = iter.next();
-
-        // Figure out which record this is and test accordingly.
-
-        RowSetReader reader = rowSet.reader();
-        assertTrue(reader.next());
-        String col2 = reader.scalar(1).getString();
-        if (col2.equals("10")) {
-          RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-              .addRow(null, "10", "foo", "bar")
-              .build();
-          RowSetUtilities.verify(expected, rowSet);
-        } else {
-          RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-              .addRow(NESTED_DIR, "20", "fred", "wilma")
-              .build();
-          RowSetUtilities.verify(expected, rowSet);
-        }
-      }
-      assertFalse(iter.hasNext());
-    } finally {
-      resetV3();
-    }
+    TupleMetadata expectedSchema = new SchemaBuilder()
+        .add("a", MinorType.VARCHAR)
+        .add("b", MinorType.VARCHAR)
+        .add("c", MinorType.VARCHAR)
+        .buildSchema();
+    RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
+        .addRow("10", "dino", "")
+        .addRow("20", "foo", "bar")
+        .addRow("30", "", "")
+        .build();
+    RowSetUtilities.verify(expected, actual);
   }
 
   /**
-   * Test partition expansion in V3.
+   * Test partition expansion .
    * <p>
    * This test is tricky because it will return two data batches
    * (preceded by an empty schema batch.) File read order is random
@@ -609,105 +343,48 @@ public class TestCsvWithHeaders extends BaseCsvTest {
    * files are nested to another level.)
    */
   @Test
-  public void testPartitionExpansionV3() throws IOException {
-    try {
-      enableV3(true);
+  public void testPartitionExpansion() throws IOException {
+    String sql = "SELECT * FROM `dfs.data`.`%s`";
+    Iterator<DirectRowSet> iter = client.queryBuilder().sql(sql, PART_DIR).rowSetIterator();
 
-      String sql = "SELECT * FROM `dfs.data`.`%s`";
-      Iterator<DirectRowSet> iter = client.queryBuilder().sql(sql, PART_DIR).rowSetIterator();
+    TupleMetadata expectedSchema = new SchemaBuilder()
+        .add("a", MinorType.VARCHAR)
+        .add("b", MinorType.VARCHAR)
+        .add("c", MinorType.VARCHAR)
+        .addNullable("dir0", MinorType.VARCHAR)
+        .buildSchema();
 
-      TupleMetadata expectedSchema = new SchemaBuilder()
-          .add("a", MinorType.VARCHAR)
-          .add("b", MinorType.VARCHAR)
-          .add("c", MinorType.VARCHAR)
-          .addNullable("dir0", MinorType.VARCHAR)
-          .buildSchema();
+    // First batch is empty; just carries the schema.
 
-      // First batch is empty; just carries the schema.
+    assertTrue(iter.hasNext());
+    RowSet rowSet = iter.next();
+    assertEquals(0, rowSet.rowCount());
+    rowSet.clear();
 
-      assertTrue(iter.hasNext());
-      RowSet rowSet = iter.next();
-      assertEquals(0, rowSet.rowCount());
-      rowSet.clear();
-
-      // Read the other two batches.
-
-      for (int i = 0; i < 2; i++) {
-        assertTrue(iter.hasNext());
-        rowSet = iter.next();
-
-        // Figure out which record this is and test accordingly.
-
-        RowSetReader reader = rowSet.reader();
-        assertTrue(reader.next());
-        String col1 = reader.scalar(0).getString();
-        if (col1.equals("10")) {
-          RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-              .addRow("10", "foo", "bar", null)
-              .build();
-          RowSetUtilities.verify(expected, rowSet);
-        } else {
-          RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-              .addRow("20", "fred", "wilma", NESTED_DIR)
-              .build();
-          RowSetUtilities.verify(expected, rowSet);
-        }
-      }
-      assertFalse(iter.hasNext());
-    } finally {
-      resetV3();
-    }
-  }
+    // Read the other two batches.
 
-  /**
-   * Test the use of partition columns with the wildcard. This works for file
-   * metadata columns, but confuses the project operator when used for
-   * partition columns. DRILL-7080.
-   */
-  @Test
-  public void testWilcardAndPartitionsMultiFilesV2() throws IOException {
-    try {
-      enableV3(false);
-
-      String sql = "SELECT *, dir0, dir1 FROM `dfs.data`.`%s`";
-      Iterator<DirectRowSet> iter = client.queryBuilder().sql(sql, PART_DIR).rowSetIterator();
-
-      TupleMetadata expectedSchema = new SchemaBuilder()
-          .addNullable("dir0", MinorType.VARCHAR)
-          .add("a", MinorType.VARCHAR)
-          .add("b", MinorType.VARCHAR)
-          .add("c", MinorType.VARCHAR)
-          .addNullable("dir00", MinorType.VARCHAR)
-          .addNullable("dir1", MinorType.INT)
-          .buildSchema();
-
-      // Read the two batches.
-
-      for (int i = 0; i < 2; i++) {
-        assertTrue(iter.hasNext());
-        RowSet rowSet = iter.next();
-
-        // Figure out which record this is and test accordingly.
-
-        RowSetReader reader = rowSet.reader();
-        assertTrue(reader.next());
-        String aCol = reader.scalar("a").getString();
-        if (aCol.equals("10")) {
-          RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-              .addRow(null, "10", "foo", "bar", null, null)
-              .build();
-          RowSetUtilities.verify(expected, rowSet);
-        } else {
-          RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-              .addRow(NESTED_DIR, "20", "fred", "wilma", NESTED_DIR, null)
-              .build();
-          RowSetUtilities.verify(expected, rowSet);
-        }
+    for (int i = 0; i < 2; i++) {
+      assertTrue(iter.hasNext());
+      rowSet = iter.next();
+
+      // Figure out which record this is and test accordingly.
+
+      RowSetReader reader = rowSet.reader();
+      assertTrue(reader.next());
+      String col1 = reader.scalar(0).getString();
+      if (col1.equals("10")) {
+        RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
+            .addRow("10", "foo", "bar", null)
+            .build();
+        RowSetUtilities.verify(expected, rowSet);
+      } else {
+        RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
+            .addRow("20", "fred", "wilma", NESTED_DIR)
+            .build();
+        RowSetUtilities.verify(expected, rowSet);
       }
-      assertFalse(iter.hasNext());
-    } finally {
-      resetV3();
     }
+    assertFalse(iter.hasNext());
   }
 
   /**
@@ -718,167 +395,103 @@ public class TestCsvWithHeaders extends BaseCsvTest {
    * partition column moves after data columns.
    */
   @Test
-  public void testWilcardAndPartitionsMultiFilesV3() throws IOException {
-    try {
-      enableV3(true);
+  public void testWilcardAndPartitionsMultiFiles() throws IOException {
+    String sql = "SELECT *, dir0, dir1 FROM `dfs.data`.`%s`";
+    Iterator<DirectRowSet> iter = client.queryBuilder().sql(sql, PART_DIR).rowSetIterator();
+
+    TupleMetadata expectedSchema = new SchemaBuilder()
+        .add("a", MinorType.VARCHAR)
+        .add("b", MinorType.VARCHAR)
+        .add("c", MinorType.VARCHAR)
+        .addNullable("dir0", MinorType.VARCHAR)
+        .addNullable("dir1", MinorType.VARCHAR)
+        .addNullable("dir00", MinorType.VARCHAR)
+        .addNullable("dir10", MinorType.VARCHAR)
+        .buildSchema();
 
-      String sql = "SELECT *, dir0, dir1 FROM `dfs.data`.`%s`";
-      Iterator<DirectRowSet> iter = client.queryBuilder().sql(sql, PART_DIR).rowSetIterator();
+    // First batch is empty; just carries the schema.
 
-      TupleMetadata expectedSchema = new SchemaBuilder()
-          .add("a", MinorType.VARCHAR)
-          .add("b", MinorType.VARCHAR)
-          .add("c", MinorType.VARCHAR)
-          .addNullable("dir0", MinorType.VARCHAR)
-          .addNullable("dir1", MinorType.VARCHAR)
-          .addNullable("dir00", MinorType.VARCHAR)
-          .addNullable("dir10", MinorType.VARCHAR)
-          .buildSchema();
+    assertTrue(iter.hasNext());
+    RowSet rowSet = iter.next();
+    RowSetUtilities.verify(new RowSetBuilder(client.allocator(), expectedSchema).build(),
+        rowSet);
 
-      // First batch is empty; just carries the schema.
+    // Read the two batches.
 
+    for (int i = 0; i < 2; i++) {
       assertTrue(iter.hasNext());
-      RowSet rowSet = iter.next();
-      RowSetUtilities.verify(new RowSetBuilder(client.allocator(), expectedSchema).build(),
-          rowSet);
-
-      // Read the two batches.
-
-      for (int i = 0; i < 2; i++) {
-        assertTrue(iter.hasNext());
-        rowSet = iter.next();
-
-        // Figure out which record this is and test accordingly.
-
-        RowSetReader reader = rowSet.reader();
-        assertTrue(reader.next());
-        String aCol = reader.scalar("a").getString();
-        if (aCol.equals("10")) {
-          RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-              .addRow("10", "foo", "bar", null, null, null, null)
-              .build();
-          RowSetUtilities.verify(expected, rowSet);
-        } else {
-          RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-              .addRow("20", "fred", "wilma", NESTED_DIR, null, NESTED_DIR, null)
-              .build();
-          RowSetUtilities.verify(expected, rowSet);
-        }
-      }
-      assertFalse(iter.hasNext());
-    } finally {
-      resetV3();
-    }
-  }
-
-  /**
-   * Test using partition columns with partitioned files in V2. Since the
-   * file is nested to one level, dir0 is a nullable VARCHAR, but dir1 is
-   * a nullable INT. Since both files are read in a single scan operator,
-   * the schema is consistent.
-   */
-  @Test
-  public void doTestExplicitPartitionsMultiFilesV2() throws IOException {
-    try {
-      enableV3(false);
-
-      String sql = "SELECT a, b, c, dir0, dir1 FROM `dfs.data`.`%s`";
-      Iterator<DirectRowSet> iter = client.queryBuilder().sql(sql, PART_DIR).rowSetIterator();
-
-      TupleMetadata expectedSchema = new SchemaBuilder()
-          .add("a", MinorType.VARCHAR)
-          .add("b", MinorType.VARCHAR)
-          .add("c", MinorType.VARCHAR)
-          .addNullable("dir0", MinorType.VARCHAR)
-          .addNullable("dir1", MinorType.INT)
-          .buildSchema();
-
-      // Read the two batches.
-
-      for (int i = 0; i < 2; i++) {
-        assertTrue(iter.hasNext());
-        RowSet rowSet = iter.next();
-
-        // Figure out which record this is and test accordingly.
-
-        RowSetReader reader = rowSet.reader();
-        assertTrue(reader.next());
-        String aCol = reader.scalar("a").getString();
-        if (aCol.equals("10")) {
-          RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-              .addRow("10", "foo", "bar", null, null)
-              .build();
-          RowSetUtilities.verify(expected, rowSet);
-        } else {
-          RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-              .addRow("20", "fred", "wilma", NESTED_DIR, null)
-              .build();
-          RowSetUtilities.verify(expected, rowSet);
-        }
+      rowSet = iter.next();
+
+      // Figure out which record this is and test accordingly.
+
+      RowSetReader reader = rowSet.reader();
+      assertTrue(reader.next());
+      String aCol = reader.scalar("a").getString();
+      if (aCol.equals("10")) {
+        RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
+            .addRow("10", "foo", "bar", null, null, null, null)
+            .build();
+        RowSetUtilities.verify(expected, rowSet);
+      } else {
+        RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
+            .addRow("20", "fred", "wilma", NESTED_DIR, null, NESTED_DIR, null)
+            .build();
+        RowSetUtilities.verify(expected, rowSet);
       }
-      assertFalse(iter.hasNext());
-    } finally {
-      resetV3();
     }
+    assertFalse(iter.hasNext());
   }
 
-  /**
+   /**
    * Test using partition columns with partitioned files in V3. Although the
    * file is nested to one level, both dir0 and dir1 are nullable VARCHAR.
    * See {@link TestPartitionRace} to show that the types and schemas
    * are consistent even when used across multiple scans.
    */
   @Test
-  public void doTestExplicitPartitionsMultiFilesV3() throws IOException {
-    try {
-      enableV3(true);
+  public void doTestExplicitPartitionsMultiFiles() throws IOException {
+    String sql = "SELECT a, b, c, dir0, dir1 FROM `dfs.data`.`%s`";
+    Iterator<DirectRowSet> iter = client.queryBuilder().sql(sql, PART_DIR).rowSetIterator();
+
+    TupleMetadata expectedSchema = new SchemaBuilder()
+        .add("a", MinorType.VARCHAR)
+        .add("b", MinorType.VARCHAR)
+        .add("c", MinorType.VARCHAR)
+        .addNullable("dir0", MinorType.VARCHAR)
+        .addNullable("dir1", MinorType.VARCHAR)
+        .buildSchema();
 
-      String sql = "SELECT a, b, c, dir0, dir1 FROM `dfs.data`.`%s`";
-      Iterator<DirectRowSet> iter = client.queryBuilder().sql(sql, PART_DIR).rowSetIterator();
+    // First batch is empty; just carries the schema.
 
-      TupleMetadata expectedSchema = new SchemaBuilder()
-          .add("a", MinorType.VARCHAR)
-          .add("b", MinorType.VARCHAR)
-          .add("c", MinorType.VARCHAR)
-          .addNullable("dir0", MinorType.VARCHAR)
-          .addNullable("dir1", MinorType.VARCHAR)
-          .buildSchema();
+    assertTrue(iter.hasNext());
+    RowSet rowSet = iter.next();
+    RowSetUtilities.verify(new RowSetBuilder(client.allocator(), expectedSchema).build(),
+        rowSet);
 
-      // First batch is empty; just carries the schema.
+    // Read the two batches.
 
+    for (int i = 0; i < 2; i++) {
       assertTrue(iter.hasNext());
-      RowSet rowSet = iter.next();
-      RowSetUtilities.verify(new RowSetBuilder(client.allocator(), expectedSchema).build(),
-          rowSet);
-
-      // Read the two batches.
-
-      for (int i = 0; i < 2; i++) {
-        assertTrue(iter.hasNext());
-        rowSet = iter.next();
-
-        // Figure out which record this is and test accordingly.
-
-        RowSetReader reader = rowSet.reader();
-        assertTrue(reader.next());
-        String aCol = reader.scalar("a").getString();
-        if (aCol.equals("10")) {
-          RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-              .addRow("10", "foo", "bar", null, null)
-              .build();
-          RowSetUtilities.verify(expected, rowSet);
-        } else {
-          RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-              .addRow("20", "fred", "wilma", NESTED_DIR, null)
-              .build();
-          RowSetUtilities.verify(expected, rowSet);
-        }
+      rowSet = iter.next();
+
+      // Figure out which record this is and test accordingly.
+
+      RowSetReader reader = rowSet.reader();
+      assertTrue(reader.next());
+      String aCol = reader.scalar("a").getString();
+      if (aCol.equals("10")) {
+        RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
+            .addRow("10", "foo", "bar", null, null)
+            .build();
+        RowSetUtilities.verify(expected, rowSet);
+      } else {
+        RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
+            .addRow("20", "fred", "wilma", NESTED_DIR, null)
+            .build();
+        RowSetUtilities.verify(expected, rowSet);
       }
-      assertFalse(iter.hasNext());
-    }
-    finally {
-      resetV3();
     }
+    assertFalse(iter.hasNext());
   }
 
   /**
@@ -886,38 +499,31 @@ public class TestCsvWithHeaders extends BaseCsvTest {
    * column when using column headers.
    */
   @Test
-  public void testColumnsColV3() throws IOException {
-    try {
-      enableV3(true);
-
-      String sql = "SELECT author, columns FROM `dfs.data`.`%s`";
-      RowSet actual = client.queryBuilder().sql(sql, COLUMNS_FILE_NAME).rowSet();
-
-      TupleMetadata expectedSchema = new SchemaBuilder()
-          .add("author", MinorType.VARCHAR)
-          .add("columns", MinorType.VARCHAR)
-          .buildSchema();
-
-      RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-          .addRow("fred", "Rocks Today,Dino Wrangling")
-          .addRow("barney", "Bowlarama")
-          .build();
-      RowSetUtilities.verify(expected, actual);
-    } finally {
-      resetV3();
-    }
+  public void testColumnsCol() throws IOException {
+    String sql = "SELECT author, columns FROM `dfs.data`.`%s`";
+    RowSet actual = client.queryBuilder().sql(sql, COLUMNS_FILE_NAME).rowSet();
+
+    TupleMetadata expectedSchema = new SchemaBuilder()
+        .add("author", MinorType.VARCHAR)
+        .add("columns", MinorType.VARCHAR)
+        .buildSchema();
+
+    RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
+        .addRow("fred", "Rocks Today,Dino Wrangling")
+        .addRow("barney", "Bowlarama")
+        .build();
+    RowSetUtilities.verify(expected, actual);
   }
 
   /**
    * The column name `columns` is treated as a plain old
    * column when using column headers. If used with an index,
    * validation will fail because the VarChar column is not an array
+   * @throws Exception
    */
   @Test
-  public void testColumnsIndexV3() throws IOException {
+  public void testColumnsIndex() throws Exception {
     try {
-      enableV3(true);
-
       String sql = "SELECT author, columns[0] FROM `dfs.data`.`%s`";
       client.queryBuilder().sql(sql, COLUMNS_FILE_NAME).run();
     } catch (UserRemoteException e) {
@@ -927,43 +533,32 @@ public class TestCsvWithHeaders extends BaseCsvTest {
       assertTrue(e.getMessage().contains("Plugin config name: csv"));
       assertTrue(e.getMessage().contains("Extract headers: true"));
       assertTrue(e.getMessage().contains("Skip first line: false"));
-    } catch (Exception e) {
-      fail();
-    } finally {
-      resetV3();
     }
   }
 
   @Test
-  public void testColumnsMissingV3() throws IOException {
-    try {
-      enableV3(true);
-
-      String sql = "SELECT a, columns FROM `dfs.data`.`%s`";
-      RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
-
-      TupleMetadata expectedSchema = new SchemaBuilder()
-          .add("a", MinorType.VARCHAR)
-          .add("columns", MinorType.VARCHAR)
-          .buildSchema();
-      RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-          .addRow("10", "")
-          .build();
-      RowSetUtilities.verify(expected, actual);
-    } finally {
-      resetV3();
-    }
+  public void testColumnsMissing() throws IOException {
+    String sql = "SELECT a, columns FROM `dfs.data`.`%s`";
+    RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
+
+    TupleMetadata expectedSchema = new SchemaBuilder()
+        .add("a", MinorType.VARCHAR)
+        .add("columns", MinorType.VARCHAR)
+        .buildSchema();
+    RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
+        .addRow("10", "")
+        .build();
+    RowSetUtilities.verify(expected, actual);
   }
 
   /**
    * If columns[x] is used, then this can't possibly match a valid
    * text reader column, so raise an error instead.
+   * @throws Exception
    */
   @Test
-  public void testColumnsIndexMissingV3() throws IOException {
+  public void testColumnsIndexMissing() throws Exception {
     try {
-      enableV3(true);
-
       String sql = "SELECT a, columns[0] FROM `dfs.data`.`%s`";
       client.queryBuilder().sql(sql, TEST_FILE_NAME).run();
     } catch (UserRemoteException e) {
@@ -975,35 +570,26 @@ public class TestCsvWithHeaders extends BaseCsvTest {
       assertTrue(e.getMessage().contains("Plugin config name: csv"));
       assertTrue(e.getMessage().contains("Extract headers: true"));
       assertTrue(e.getMessage().contains("Skip first line: false"));
-    } catch (Exception e) {
-      fail();
-    } finally {
-      resetV3();
     }
   }
 
   @Test
   public void testHugeColumn() throws IOException {
     String fileName = buildBigColFile(true);
-    try {
-      enableV3(true);
-      String sql = "SELECT * FROM `dfs.data`.`%s`";
-      RowSet actual = client.queryBuilder().sql(sql, fileName).rowSet();
-      assertEquals(10, actual.rowCount());
-      RowSetReader reader = actual.reader();
-      while (reader.next()) {
-        int i = reader.logicalIndex();
-        assertEquals(Integer.toString(i + 1), reader.scalar(0).getString());
-        String big = reader.scalar(1).getString();
-        assertEquals(BIG_COL_SIZE, big.length());
-        for (int j = 0; j < BIG_COL_SIZE; j++) {
-          assertEquals((char) ((j + i) % 26 + 'A'), big.charAt(j));
-        }
-        assertEquals(Integer.toString((i + 1) * 10), reader.scalar(2).getString());
+    String sql = "SELECT * FROM `dfs.data`.`%s`";
+    RowSet actual = client.queryBuilder().sql(sql, fileName).rowSet();
+    assertEquals(10, actual.rowCount());
+    RowSetReader reader = actual.reader();
+    while (reader.next()) {
+      int i = reader.logicalIndex();
+      assertEquals(Integer.toString(i + 1), reader.scalar(0).getString());
+      String big = reader.scalar(1).getString();
+      assertEquals(BIG_COL_SIZE, big.length());
+      for (int j = 0; j < BIG_COL_SIZE; j++) {
+        assertEquals((char) ((j + i) % 26 + 'A'), big.charAt(j));
       }
-      actual.clear();
-    } finally {
-      resetV3();
+      assertEquals(Integer.toString((i + 1) * 10), reader.scalar(2).getString());
     }
+    actual.clear();
   }
 }
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvWithSchema.java b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvWithSchema.java
index 955eb3d..9a76c1f 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvWithSchema.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvWithSchema.java
@@ -34,14 +34,12 @@ import org.apache.drill.test.rowSet.DirectRowSet;
 import org.apache.drill.test.rowSet.RowSet;
 import org.apache.drill.test.rowSet.RowSetBuilder;
 import org.apache.drill.test.rowSet.RowSetComparison;
-import org.apache.drill.test.rowSet.RowSetReader;
 import org.apache.drill.test.rowSet.RowSetUtilities;
 import org.joda.time.Instant;
 import org.joda.time.LocalDate;
 import org.joda.time.LocalTime;
 import org.joda.time.Period;
 import org.junit.BeforeClass;
-import org.junit.Ignore;
 import org.junit.Test;
 import org.junit.experimental.categories.Category;
 
@@ -74,11 +72,6 @@ public class TestCsvWithSchema extends BaseCsvTest {
     "4,betty,2019-05-04,NA"
   };
 
-  private static final String multi2FullContents[] = {
-    "id,name,date",
-    "3,barney,2001-01-16,NA"
-  };
-
   private static final String reordered2Contents[] = {
     "name,id,date",
     "barney,3,2001-01-16"
@@ -308,93 +301,6 @@ public class TestCsvWithSchema extends BaseCsvTest {
   }
 
   /**
-   * Test the schema we get in V2 when the table read order is random.
-   * Worst-case: the two files have different column counts and
-   * column orders.
-   * <p>
-   * Though the results are random, we iterate 10 times which, in most runs,
-   * shows the random variation in schemas:
-   * <ul>
-   * <li>Sometimes the first batch has three columns, sometimes four.</li>
-   * <li>Sometimes the column `id` is in position 0, sometimes in position 1
-   * (correlated with the above).</li>
-   * <li>Due to the fact that sometimes the first file (with four columns)
-   * is returned first, sometimes the second file (with three columns) is
-   * returned first.</li>
-   * </ul>
-   */
-  @Test
-  public void testSchemaRaceV2() throws Exception {
-    try {
-      enableV3(false);
-      enableSchema(false);
-      enableMultiScan();
-      String tablePath = buildTable("schemaRaceV2", multi1Contents, reordered2Contents);
-      boolean sawFile1First = false;
-      boolean sawFile2First = false;
-      boolean sawFullSchema = false;
-      boolean sawPartialSchema = false;
-      boolean sawIdAsCol0 = false;
-      boolean sawIdAsCol1 = false;
-      String sql = "SELECT * FROM " + tablePath;
-      for (int i = 0; i < 10; i++) {
-        Iterator<DirectRowSet> iter = client.queryBuilder().sql(sql).rowSetIterator();
-        int batchCount = 0;
-        while(iter.hasNext()) {
-          batchCount++;
-          RowSet result = iter.next();
-          TupleMetadata resultSchema = result.schema();
-          if (resultSchema.size() == 4) {
-            sawFullSchema = true;
-          } else {
-            assertEquals(3, resultSchema.size());
-            sawPartialSchema = true;
-          }
-          if (resultSchema.index("id") == 0) {
-            sawIdAsCol0 = true;
-          } else {
-            assertEquals(1, resultSchema.index("id"));
-            sawIdAsCol1 = true;
-          }
-          if (batchCount == 1) {
-            RowSetReader reader = result.reader();
-            assertTrue(reader.next());
-            String id = reader.scalar("id").getString();
-            if (id.equals("1")) {
-              sawFile1First = true;
-            } else {
-              assertEquals("3", id);
-              sawFile2First = true;
-            }
-          }
-          result.clear();
-        }
-      }
-
-      // Outcome is random (which is the key problem). Don't assert on these
-      // because doing so can lead to a flakey test.
-
-      if (!sawFile1First || ! sawFile2First || !sawFullSchema || !sawPartialSchema || !sawIdAsCol0 || !sawIdAsCol1) {
-        System.out.println("Some variations did not occur");
-        System.out.println(String.format("File 1 first: %s", sawFile1First));
-        System.out.println(String.format("File 1 second: %s", sawFile2First));
-        System.out.println(String.format("Full schema: %s", sawFullSchema));
-        System.out.println(String.format("Partial schema: %s", sawPartialSchema));
-        System.out.println(String.format("`id` as col 0: %s", sawIdAsCol0));
-        System.out.println(String.format("`id` as col 1: %s", sawIdAsCol1));
-      }
-      // Sanity checks
-      assertTrue(sawFullSchema);
-      assertTrue(sawFile1First || sawFile2First);
-      assertTrue(sawIdAsCol0 || sawIdAsCol1);
-    } finally {
-      resetV3();
-      resetSchema();
-      resetMultiScan();
-    }
-  }
-
-  /**
    * Show that, without schema, the hard schema change for the "missing"
    * gender column causes an error in the sort operator when presented with
    * one batch in which gender is VARCHAR, another in which it is
@@ -406,89 +312,23 @@ public class TestCsvWithSchema extends BaseCsvTest {
     try {
       enableSchema(false);
       enableMultiScan();
-      enableV3(false);
       String tablePath = buildTable("wildcardSortV2", multi1Contents, reordered2Contents);
-      doTestWildcardSortFailure(tablePath);
-      enableV3(true);
-      doTestWildcardSortFailure(tablePath);
-    } finally {
-      resetV3();
-      resetSchema();
-      resetMultiScan();
-    }
-  }
-
-  private void doTestWildcardSortFailure(String tablePath) throws Exception {
-    String sql = "SELECT * FROM " + tablePath + " ORDER BY id";
-    boolean sawError = false;
-    for (int i = 0; i < 10; i++) {
-      try {
-        // When this fails it will print a nasty stack trace.
-        RowSet result = client.queryBuilder().sql(sql).rowSet();
-        assertEquals(4, result.rowCount());
-        result.clear();
-      } catch (RpcException e) {
-        assertTrue(e.getCause() instanceof UserRemoteException);
-        sawError = true;
-        break;
-      }
-    }
-    assertTrue(sawError);
-  }
-
-  /**
-   * Test an explicit projection with a sort. Using the sort 1) will blow up
-   * if the internal schema is inconsistent, and 2) allows easier verification
-   * of the merged table results.
-   * <p>
-   * Fails with <code><pre>
-   * #: id, name, gender
-   * 0: "1", "barney", ""
-   * 1: "2", "", ""
-   * 2: "3", "
-   * ", ""
-   * 3: "4", "
-   * " java.lang.NegativeArraySizeException: null
-   *   at io.netty.buffer.DrillBuf.unsafeGetMemory(DrillBuf.java:852) ~[classes/:4.0.48.Final]
-   * </pre></code>
-   */
-  @Test
-  @Ignore("Vectors get corrupted somehow")
-  public void testV2ExplicitSortFailure() throws Exception {
-    try {
-      enableSchema(false);
-      enableMultiScan();
-      enableV3(false);
-      // V2 fails on ragged columns, use consistent columns
-      String tablePath = buildTable("explicitSortV2", multi1Contents, multi2FullContents);
-      String sql = "SELECT id, name, gender FROM " + tablePath + " ORDER BY id";
-      TupleMetadata expectedSchema = new SchemaBuilder()
-          .add("id", MinorType.VARCHAR)
-          .add("name", MinorType.VARCHAR)
-          .add("gender", MinorType.VARCHAR)
-          .buildSchema();
-      RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-          .addRow("1", "wilma", "female")
-          .addRow("2", "fred", "male")
-          .addRow("3", "barney", "NA")
-          .addRow("4", "betty", "NA")
-          .build();
+      String sql = "SELECT * FROM " + tablePath + " ORDER BY id";
       boolean sawError = false;
       for (int i = 0; i < 10; i++) {
         try {
+          // When this fails it will print a nasty stack trace.
           RowSet result = client.queryBuilder().sql(sql).rowSet();
-          result.print();
-          new RowSetComparison(expected).verifyAndClear(result);
+          assertEquals(4, result.rowCount());
+          result.clear();
         } catch (RpcException e) {
           assertTrue(e.getCause() instanceof UserRemoteException);
           sawError = true;
           break;
         }
       }
-      expected.clear();
       assertTrue(sawError);
     } finally {
-      resetV3();
       resetSchema();
       resetMultiScan();
     }
@@ -504,13 +344,11 @@ public class TestCsvWithSchema extends BaseCsvTest {
    * orders (file1 first sometimes, file2 other times.)
    */
   @Test
-  public void testV3ExplicitSort() throws Exception {
+  public void testExplicitSort() throws Exception {
     try {
       enableSchema(false);
       enableMultiScan();
-      enableV3(true);
-      // V3 handles ragged columns
-      String tablePath = buildTable("v3ExplictSort", raggedMulti1Contents, reordered2Contents);
+      String tablePath = buildTable("explictSort1", raggedMulti1Contents, reordered2Contents);
       String sql = "SELECT id, name, gender FROM " + tablePath + " ORDER BY id";
       TupleMetadata expectedSchema = new SchemaBuilder()
           .add("id", MinorType.VARCHAR)
@@ -529,7 +367,6 @@ public class TestCsvWithSchema extends BaseCsvTest {
       }
       expected.clear();
     } finally {
-      resetV3();
       resetSchema();
       resetMultiScan();
     }
@@ -547,7 +384,7 @@ public class TestCsvWithSchema extends BaseCsvTest {
       enableSchemaSupport();
       enableMultiScan();
       // V3 handles ragged columns
-      String tablePath = buildTable("v3ExplictSort", raggedMulti1Contents, reordered2Contents);
+      String tablePath = buildTable("explictSort2", raggedMulti1Contents, reordered2Contents);
       run(SCHEMA_SQL, tablePath);
       String sql = "SELECT id, name, gender FROM " + tablePath + " ORDER BY id";
       TupleMetadata expectedSchema = new SchemaBuilder()
@@ -641,8 +478,8 @@ public class TestCsvWithSchema extends BaseCsvTest {
    * variation in inputs.
    */
   @Test
-  public void testWildcardV3LenientSchema() throws Exception {
-    String tableName = "wildcardLenientV3";
+  public void testWildcardLenientSchema() throws Exception {
+    String tableName = "wildcardLenient";
     String tablePath = buildTable(tableName, multi1Contents,
         reordered2Contents, nameOnlyContents);
 
@@ -677,8 +514,8 @@ public class TestCsvWithSchema extends BaseCsvTest {
    * projected.
    */
   @Test
-  public void testWildcardV3StrictSchema() throws Exception {
-    String tableName = "wildcardStrictV3";
+  public void testWildcardStrictSchema() throws Exception {
+    String tableName = "wildcardStrict";
     String tablePath = buildTable(tableName, multi1Contents,
         reordered2Contents, nameOnlyContents);
 
@@ -717,7 +554,7 @@ public class TestCsvWithSchema extends BaseCsvTest {
    */
   @Test
   public void testMultiFragmentStrictSchema() throws Exception {
-    String tableName = "wildcardStrict2V3";
+    String tableName = "wildcardStrict2";
     String tablePath = buildTable(tableName, multi1Contents,
         reordered2Contents, nameOnlyContents);
 
@@ -1114,7 +951,6 @@ public class TestCsvWithSchema extends BaseCsvTest {
           .build();
       RowSetUtilities.verify(expected, actual);
     } finally {
-      resetV3();
       resetSchema();
     }
   }
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvWithoutHeaders.java b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvWithoutHeaders.java
index 1340d90..8918b39 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvWithoutHeaders.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestCsvWithoutHeaders.java
@@ -21,7 +21,6 @@ import static org.apache.drill.test.rowSet.RowSetUtilities.strArray;
 import static org.junit.Assert.assertEquals;
 import static org.junit.Assert.assertFalse;
 import static org.junit.Assert.assertTrue;
-import static org.junit.Assert.fail;
 
 import java.io.File;
 import java.io.IOException;
@@ -42,7 +41,12 @@ import org.junit.BeforeClass;
 import org.junit.Test;
 import org.junit.experimental.categories.Category;
 
-// CSV reader now hosted on the row set framework
+/**
+ * Test behavior of the text (CSV) reader for files without headers
+ * and without an external schema. Data is represented with the
+ * `columns` array column.
+ */
+
 @Category(RowSetTests.class)
 public class TestCsvWithoutHeaders extends BaseCsvTest {
 
@@ -83,23 +87,12 @@ public class TestCsvWithoutHeaders extends BaseCsvTest {
     buildFile(new File(nestedDir, NESTED_FILE), secondSet);
   }
 
-  @Test
-  public void testWildcard() throws IOException {
-    try {
-      enableV3(false);
-      doTestWildcard();
-      enableV3(true);
-      doTestWildcard();
-    } finally {
-      resetV3();
-    }
-  }
-
   /**
    * Verify that the wildcard expands to the `columns` array
    */
 
-  private void doTestWildcard() throws IOException {
+  @Test
+  public void testWildcard() throws IOException {
     String sql = "SELECT * FROM `dfs.data`.`%s`";
     RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
 
@@ -116,17 +109,6 @@ public class TestCsvWithoutHeaders extends BaseCsvTest {
 
   @Test
   public void testColumns() throws IOException {
-    try {
-      enableV3(false);
-      doTestColumns();
-      enableV3(true);
-      doTestColumns();
-    } finally {
-      resetV3();
-    }
-  }
-
-  private void doTestColumns() throws IOException {
     String sql = "SELECT columns FROM `dfs.data`.`%s`";
     RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
 
@@ -142,106 +124,41 @@ public class TestCsvWithoutHeaders extends BaseCsvTest {
   }
 
   @Test
-  public void doTestWildcardAndMetadataV2() throws IOException {
-    try {
-      enableV3(false);
-      String sql = "SELECT *, filename FROM `dfs.data`.`%s`";
-      RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
-
-      TupleMetadata expectedSchema = new SchemaBuilder()
-          .addArray("columns", MinorType.VARCHAR)
-          .addNullable("filename", MinorType.VARCHAR)
-          .buildSchema();
-
-      RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-          .addRow(strArray("10", "foo", "bar"), TEST_FILE_NAME)
-          .addRow(strArray("20", "fred", "wilma"), TEST_FILE_NAME)
-          .build();
-      RowSetUtilities.verify(expected, actual);
-    } finally {
-      resetV3();
-    }
-  }
+  public void doTestWildcardAndMetadata() throws IOException {
+    String sql = "SELECT *, filename FROM `dfs.data`.`%s`";
+    RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
 
-  @Test
-  public void doTestWildcardAndMetadataV3() throws IOException {
-    try {
-      enableV3(true);
-      String sql = "SELECT *, filename FROM `dfs.data`.`%s`";
-      RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
-
-      TupleMetadata expectedSchema = new SchemaBuilder()
-          .addArray("columns", MinorType.VARCHAR)
-          .add("filename", MinorType.VARCHAR)
-          .buildSchema();
-
-      RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-          .addRow(strArray("10", "foo", "bar"), TEST_FILE_NAME)
-          .addRow(strArray("20", "fred", "wilma"), TEST_FILE_NAME)
-          .build();
-      RowSetUtilities.verify(expected, actual);
-    } finally {
-      resetV3();
-    }
-  }
+    TupleMetadata expectedSchema = new SchemaBuilder()
+        .addArray("columns", MinorType.VARCHAR)
+        .add("filename", MinorType.VARCHAR)
+        .buildSchema();
 
-  @Test
-  public void testColumnsAndMetadataV2() throws IOException {
-    try {
-      enableV3(false);
-      String sql = "SELECT columns, filename FROM `dfs.data`.`%s`";
-      RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
-
-      TupleMetadata expectedSchema = new SchemaBuilder()
-          .addArray("columns", MinorType.VARCHAR)
-          .addNullable("filename", MinorType.VARCHAR)
-          .buildSchema();
-
-      RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-          .addRow(strArray("10", "foo", "bar"), TEST_FILE_NAME)
-          .addRow(strArray("20", "fred", "wilma"), TEST_FILE_NAME)
-          .build();
-      RowSetUtilities.verify(expected, actual);
-    } finally {
-      resetV3();
-    }
+    RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
+        .addRow(strArray("10", "foo", "bar"), TEST_FILE_NAME)
+        .addRow(strArray("20", "fred", "wilma"), TEST_FILE_NAME)
+        .build();
+    RowSetUtilities.verify(expected, actual);
   }
 
   @Test
-  public void testColumnsAndMetadataV3() throws IOException {
-    try {
-      enableV3(true);
-      String sql = "SELECT columns, filename FROM `dfs.data`.`%s`";
-      RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
-
-      TupleMetadata expectedSchema = new SchemaBuilder()
-          .addArray("columns", MinorType.VARCHAR)
-          .add("filename", MinorType.VARCHAR)
-          .buildSchema();
-
-      RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-          .addRow(strArray("10", "foo", "bar"), TEST_FILE_NAME)
-          .addRow(strArray("20", "fred", "wilma"), TEST_FILE_NAME)
-          .build();
-      RowSetUtilities.verify(expected, actual);
-    } finally {
-      resetV3();
-    }
+  public void testColumnsAndMetadata() throws IOException {
+    String sql = "SELECT columns, filename FROM `dfs.data`.`%s`";
+    RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
+
+    TupleMetadata expectedSchema = new SchemaBuilder()
+        .addArray("columns", MinorType.VARCHAR)
+        .add("filename", MinorType.VARCHAR)
+        .buildSchema();
+
+    RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
+        .addRow(strArray("10", "foo", "bar"), TEST_FILE_NAME)
+        .addRow(strArray("20", "fred", "wilma"), TEST_FILE_NAME)
+        .build();
+    RowSetUtilities.verify(expected, actual);
   }
 
   @Test
   public void testSpecificColumns() throws IOException {
-    try {
-      enableV3(false);
-      doTestSpecificColumns();
-      enableV3(true);
-      doTestSpecificColumns();
-    } finally {
-      resetV3();
-    }
-  }
-
-  private void doTestSpecificColumns() throws IOException {
     String sql = "SELECT columns[0], columns[2] FROM `dfs.data`.`%s`";
     RowSet actual = client.queryBuilder().sql(sql, TEST_FILE_NAME).rowSet();
 
@@ -261,17 +178,6 @@ public class TestCsvWithoutHeaders extends BaseCsvTest {
   public void testRaggedRows() throws IOException {
     String fileName = "ragged.csv";
     buildFile(fileName, raggedRows);
-    try {
-      enableV3(false);
-      doTestRaggedRows(fileName);
-      enableV3(true);
-      doTestRaggedRows(fileName);
-    } finally {
-      resetV3();
-    }
-  }
-
-  private void doTestRaggedRows(String fileName) throws IOException {
     String sql = "SELECT columns FROM `dfs.data`.`%s`";
     RowSet actual = client.queryBuilder().sql(sql, fileName).rowSet();
 
@@ -287,57 +193,6 @@ public class TestCsvWithoutHeaders extends BaseCsvTest {
   }
 
   /**
-   * Test partition expansion. Because the two files are read in the
-   * same scan operator, the schema is consistent.
-   * <p>
-   * V2, since Drill 1.12, puts partition columns ahead of data columns.
-   */
-  @Test
-  public void testPartitionExpansionV2() throws IOException {
-    try {
-      enableV3(false);
-
-      String sql = "SELECT * FROM `dfs.data`.`%s`";
-      Iterator<DirectRowSet> iter = client.queryBuilder().sql(sql, PART_DIR).rowSetIterator();
-
-      TupleMetadata expectedSchema = new SchemaBuilder()
-          .addNullable("dir0", MinorType.VARCHAR)
-          .addArray("columns", MinorType.VARCHAR)
-          .buildSchema();
-
-      // Read the two batches.
-
-      for (int i = 0; i < 2; i++) {
-        assertTrue(iter.hasNext());
-        RowSet rowSet = iter.next();
-
-        // Figure out which record this is and test accordingly.
-
-        RowSetReader reader = rowSet.reader();
-        assertTrue(reader.next());
-        ArrayReader ar = reader.array(1);
-        assertTrue(ar.next());
-        String col1 = ar.scalar().getString();
-        if (col1.equals("10")) {
-          RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-              .addRow(null, strArray("10", "foo", "bar"))
-              .addRow(null, strArray("20", "fred", "wilma"))
-              .build();
-          RowSetUtilities.verify(expected, rowSet);
-        } else {
-          RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-              .addRow(NESTED_DIR, strArray("30", "barney", "betty"))
-              .build();
-          RowSetUtilities.verify(expected, rowSet);
-        }
-      }
-      assertFalse(iter.hasNext());
-    } finally {
-      resetV3();
-    }
-  }
-
-  /**
    * Test partition expansion in V3.
    * <p>
    * V3, as in V2 before Drill 1.12, puts partition columns after
@@ -346,54 +201,48 @@ public class TestCsvWithoutHeaders extends BaseCsvTest {
    */
   @Test
   public void testPartitionExpansionV3() throws IOException {
-    try {
-      enableV3(true);
+    String sql = "SELECT * FROM `dfs.data`.`%s`";
+    Iterator<DirectRowSet> iter = client.queryBuilder().sql(sql, PART_DIR).rowSetIterator();
+
+    TupleMetadata expectedSchema = new SchemaBuilder()
+        .addArray("columns", MinorType.VARCHAR)
+        .addNullable("dir0", MinorType.VARCHAR)
+        .buildSchema();
 
-      String sql = "SELECT * FROM `dfs.data`.`%s`";
-      Iterator<DirectRowSet> iter = client.queryBuilder().sql(sql, PART_DIR).rowSetIterator();
+    // First batch is empty; just carries the schema.
 
-      TupleMetadata expectedSchema = new SchemaBuilder()
-          .addArray("columns", MinorType.VARCHAR)
-          .addNullable("dir0", MinorType.VARCHAR)
-          .buildSchema();
+    assertTrue(iter.hasNext());
+    RowSet rowSet = iter.next();
+    assertEquals(0, rowSet.rowCount());
+    rowSet.clear();
 
-      // First batch is empty; just carries the schema.
+    // Read the other two batches.
 
+    for (int i = 0; i < 2; i++) {
       assertTrue(iter.hasNext());
-      RowSet rowSet = iter.next();
-      assertEquals(0, rowSet.rowCount());
-      rowSet.clear();
-
-      // Read the other two batches.
-
-      for (int i = 0; i < 2; i++) {
-        assertTrue(iter.hasNext());
-        rowSet = iter.next();
-
-        // Figure out which record this is and test accordingly.
-
-        RowSetReader reader = rowSet.reader();
-        assertTrue(reader.next());
-        ArrayReader ar = reader.array(0);
-        assertTrue(ar.next());
-        String col1 = ar.scalar().getString();
-        if (col1.equals("10")) {
-          RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-              .addRow(strArray("10", "foo", "bar"), null)
-              .addRow(strArray("20", "fred", "wilma"), null)
-              .build();
-          RowSetUtilities.verify(expected, rowSet);
-        } else {
-          RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-              .addRow(strArray("30", "barney", "betty"), NESTED_DIR)
-              .build();
-          RowSetUtilities.verify(expected, rowSet);
-        }
+      rowSet = iter.next();
+
+      // Figure out which record this is and test accordingly.
+
+      RowSetReader reader = rowSet.reader();
+      assertTrue(reader.next());
+      ArrayReader ar = reader.array(0);
+      assertTrue(ar.next());
+      String col1 = ar.scalar().getString();
+      if (col1.equals("10")) {
+        RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
+            .addRow(strArray("10", "foo", "bar"), null)
+            .addRow(strArray("20", "fred", "wilma"), null)
+            .build();
+        RowSetUtilities.verify(expected, rowSet);
+      } else {
+        RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
+            .addRow(strArray("30", "barney", "betty"), NESTED_DIR)
+            .build();
+        RowSetUtilities.verify(expected, rowSet);
       }
-      assertFalse(iter.hasNext());
-    } finally {
-      resetV3();
     }
+    assertFalse(iter.hasNext());
   }
 
   /**
@@ -401,22 +250,18 @@ public class TestCsvWithoutHeaders extends BaseCsvTest {
    * implicitly suggest that `columns` is a map.
    * <p>
    * V2 message: DATA_READ ERROR: Selected column 'columns' must be an array index
+   * @throws Exception
    */
 
   @Test
-  public void testColumnsAsMap() throws IOException {
+  public void testColumnsAsMap() throws Exception {
     String sql = "SELECT `%s`.columns.foo FROM `dfs.data`.`%s`";
     try {
-      enableV3(true);
       client.queryBuilder().sql(sql, TEST_FILE_NAME, TEST_FILE_NAME).run();
     } catch (UserRemoteException e) {
       assertTrue(e.getMessage().contains(
           "VALIDATION ERROR: Column `columns` has map elements, but must be an array"));
       assertTrue(e.getMessage().contains("Plugin config name: csv"));
-    } catch (Exception e) {
-      fail();
-    } finally {
-      resetV3();
     }
   }
   /**
@@ -424,51 +269,42 @@ public class TestCsvWithoutHeaders extends BaseCsvTest {
    * it must be below the maximum.
    * <p>
    * V2 message: INTERNAL_ERROR ERROR: 70000
+   * @throws Exception
    */
 
   @Test
-  public void testColumnsIndexOverflow() throws IOException {
+  public void testColumnsIndexOverflow() throws Exception {
     String sql = "SELECT columns[70000] FROM `dfs.data`.`%s`";
     try {
-      enableV3(true);
       client.queryBuilder().sql(sql, TEST_FILE_NAME, TEST_FILE_NAME).run();
     } catch (UserRemoteException e) {
       assertTrue(e.getMessage().contains(
           "VALIDATION ERROR: `columns`[70000] index out of bounds, max supported size is 65536"));
       assertTrue(e.getMessage().contains("Plugin config name: csv"));
-    } catch (Exception e) {
-      fail();
-    } finally {
-      resetV3();
     }
   }
 
   @Test
   public void testHugeColumn() throws IOException {
     String fileName = buildBigColFile(false);
-    try {
-      enableV3(true);
-      String sql = "SELECT * FROM `dfs.data`.`%s`";
-      RowSet actual = client.queryBuilder().sql(sql, fileName).rowSet();
-      assertEquals(10, actual.rowCount());
-      RowSetReader reader = actual.reader();
-      ArrayReader arrayReader = reader.array(0);
-      while (reader.next()) {
-        int i = reader.logicalIndex();
-        arrayReader.next();
-        assertEquals(Integer.toString(i + 1), arrayReader.scalar().getString());
-        arrayReader.next();
-        String big = arrayReader.scalar().getString();
-        assertEquals(BIG_COL_SIZE, big.length());
-        for (int j = 0; j < BIG_COL_SIZE; j++) {
-          assertEquals((char) ((j + i) % 26 + 'A'), big.charAt(j));
-        }
-        arrayReader.next();
-        assertEquals(Integer.toString((i + 1) * 10), arrayReader.scalar().getString());
+    String sql = "SELECT * FROM `dfs.data`.`%s`";
+    RowSet actual = client.queryBuilder().sql(sql, fileName).rowSet();
+    assertEquals(10, actual.rowCount());
+    RowSetReader reader = actual.reader();
+    ArrayReader arrayReader = reader.array(0);
+    while (reader.next()) {
+      int i = reader.logicalIndex();
+      arrayReader.next();
+      assertEquals(Integer.toString(i + 1), arrayReader.scalar().getString());
+      arrayReader.next();
+      String big = arrayReader.scalar().getString();
+      assertEquals(BIG_COL_SIZE, big.length());
+      for (int j = 0; j < BIG_COL_SIZE; j++) {
+        assertEquals((char) ((j + i) % 26 + 'A'), big.charAt(j));
       }
-      actual.clear();
-    } finally {
-      resetV3();
+      arrayReader.next();
+      assertEquals(Integer.toString((i + 1) * 10), arrayReader.scalar().getString());
     }
+    actual.clear();
   }
 }
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestHeaderBuilder.java b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestHeaderBuilder.java
index 67429fb..42b687b 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestHeaderBuilder.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestHeaderBuilder.java
@@ -21,7 +21,7 @@ import static org.junit.Assert.assertEquals;
 import static org.junit.Assert.assertTrue;
 
 import org.apache.drill.common.exceptions.UserException;
-import org.apache.drill.exec.store.easy.text.compliant.v3.HeaderBuilder;
+import org.apache.drill.exec.store.easy.text.reader.HeaderBuilder;
 import org.apache.drill.test.DrillTest;
 import org.apache.hadoop.fs.Path;
 import org.junit.Test;
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestPartitionRace.java b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestPartitionRace.java
index 9302a9a..b1a2c94 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestPartitionRace.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestPartitionRace.java
@@ -26,7 +26,6 @@ import java.io.IOException;
 import java.util.Iterator;
 
 import org.apache.drill.common.types.TypeProtos.MinorType;
-import org.apache.drill.exec.ExecConstants;
 import org.apache.drill.exec.record.metadata.SchemaBuilder;
 import org.apache.drill.exec.record.metadata.TupleMetadata;
 import org.apache.drill.test.rowSet.DirectRowSet;
@@ -49,14 +48,8 @@ import org.junit.Test;
  * persist.
  * <p>
  * The solution is to figure out the max partition depth in the
- * EasySubScan rather than in each scan operator.
- * <p>
- * The tests here test both the "V2" (AKA "new text reader") which has
- * many issues, and the "V3" (row-set-based version) that has fixes.
- * <p>
- * See DRILL-7082 for the multi-scan race (fixed in V3), and
- * DRILL-7083 for the problem with partition columns returning nullable INT
- * (also fixed in V3.)
+ * EasySubScan rather than in each scan operator, which is done in the
+ * current "V3" version. The tests here verify this behavior.
  */
 
 public class TestPartitionRace extends BaseCsvTest {
@@ -77,61 +70,11 @@ public class TestPartitionRace extends BaseCsvTest {
 
   /**
    * Oddly, when run in a single fragment, the files occur in a
-   * stable order, the partition always appars, and it appears in
+   * stable order, the partition always appears, and it appears in
    * the first column position.
    */
   @Test
-  public void testSingleScanV2() throws IOException {
-    String sql = "SELECT * FROM `dfs.data`.`%s`";
-
-    try {
-      enableV3(false);
-
-      // Loop to run the query 10 times, or until we see the race
-
-      boolean sawMissingPartition = false;
-      boolean sawPartitionFirst = false;
-      boolean sawPartitionLast = false;
-
-      // Read the two batches.
-
-      Iterator<DirectRowSet> iter = client.queryBuilder().sql(sql, PART_DIR).rowSetIterator();
-      for (int j = 0; j < 2; j++) {
-        assertTrue(iter.hasNext());
-        RowSet rowSet = iter.next();
-
-        // Check location of partition column
-
-        int posn = rowSet.schema().index("dir0");
-        if (posn == -1) {
-          sawMissingPartition = true;
-        } else if (posn == 0) {
-          sawPartitionFirst = true;
-        } else {
-          sawPartitionLast = true;
-        }
-        rowSet.clear();
-      }
-      assertFalse(iter.hasNext());
-
-      // When run in a single fragment, the partition column appears
-      // all the time, and is in the first column position.
-
-      assertFalse(sawMissingPartition);
-      assertTrue(sawPartitionFirst);
-      assertFalse(sawPartitionLast);
-    } finally {
-      resetV3();
-      client.resetSession(ExecConstants.MIN_READER_WIDTH_KEY);
-    }
-  }
-
-  /**
-   * V3 provides the same schema for the single- and multi-scan
-   * cases.
-   */
-  @Test
-  public void testSingleScanV3() throws IOException {
+  public void testSingleScan() throws IOException {
     String sql = "SELECT * FROM `dfs.data`.`%s`";
 
     TupleMetadata expectedSchema = new SchemaBuilder()
@@ -141,142 +84,40 @@ public class TestPartitionRace extends BaseCsvTest {
         .addNullable("dir0", MinorType.VARCHAR)
         .buildSchema();
 
-    try {
-      enableV3(true);
-
-      // Loop to run the query 10 times to verify no race
-
-      // First batch is empty; just carries the schema.
-
-      Iterator<DirectRowSet> iter = client.queryBuilder().sql(sql, PART_DIR).rowSetIterator();
-      assertTrue(iter.hasNext());
-      RowSet rowSet = iter.next();
-      assertEquals(0, rowSet.rowCount());
-      rowSet.clear();
+    // Loop to run the query 10 times to verify no race
 
-      // Read the two batches.
+    // First batch is empty; just carries the schema.
 
-      for (int j = 0; j < 2; j++) {
-        assertTrue(iter.hasNext());
-        rowSet = iter.next();
+    Iterator<DirectRowSet> iter = client.queryBuilder().sql(sql, PART_DIR).rowSetIterator();
+    assertTrue(iter.hasNext());
+    RowSet rowSet = iter.next();
+    assertEquals(0, rowSet.rowCount());
+    rowSet.clear();
 
-        // Figure out which record this is and test accordingly.
+    // Read the two batches.
 
-        RowSetReader reader = rowSet.reader();
-        assertTrue(reader.next());
-        String col1 = reader.scalar("a").getString();
-        if (col1.equals("10")) {
-          RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-              .addRow("10", "foo", "bar", null)
-              .build();
-          RowSetUtilities.verify(expected, rowSet);
-        } else {
-          RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
-              .addRow("20", "fred", "wilma", NESTED_DIR)
-              .build();
-          RowSetUtilities.verify(expected, rowSet);
-        }
-      }
-      assertFalse(iter.hasNext());
-    } finally {
-      resetV3();
-    }
-  }
-
-  /**
-   * When forced to run in two fragments, the fun really starts. The
-   * partition column (usually) appears in the last column position instead
-   * of the first. The partition may or may not occur in the first row
-   * depending on which file is read first. The result is that the
-   * other columns will jump around. If we tried to create an expected
-   * result set, we'd be frustrated because the schema randomly changes.
-   * <p>
-   * Just to be clear: this behavior is a bug, not a feature. But, it is
-   * an established baseline for the "V2" reader.
-   * <p>
-   * This is really a test (demonstration) of the wrong behavior. This test
-   * is pretty unreliable. In particular, the position of the partition column
-   * seems to randomly shift from first to last position across runs.
-   */
-  @Test
-  public void testRaceV2() throws IOException {
-    String sql = "SELECT * FROM `dfs.data`.`%s`";
-
-    try {
-      enableV3(false);
-      enableMultiScan();
-
-      // Loop to run the query 10 times, or until we see the race
-
-      boolean sawRootFirst = false;
-      boolean sawNestedFirst = false;
-      boolean sawMissingPartition = false;
-      boolean sawPartitionFirst = false;
-      boolean sawPartitionLast = false;
-      for (int i = 0; i < 10; i++) {
-
-        // Read the two batches.
-
-        Iterator<DirectRowSet> iter = client.queryBuilder().sql(sql, PART_DIR).rowSetIterator();
-        for (int j = 0; j < 2; j++) {
-          assertTrue(iter.hasNext());
-          RowSet rowSet = iter.next();
-
-          // Check location of partition column
-
-          int posn = rowSet.schema().index("dir0");
-          if (posn == -1) {
-            sawMissingPartition = true;
-          } else if (posn == 0) {
-            sawPartitionFirst = true;
-          } else {
-            sawPartitionLast = true;
-          }
-
-          // Figure out which record this is and test accordingly.
-
-          RowSetReader reader = rowSet.reader();
-          assertTrue(reader.next());
-          String col1 = reader.scalar("a").getString();
-          if (col1.equals("10")) {
-            if (i == 0) {
-              sawRootFirst = true;
-            }
-          } else {
-            if (i == 0) {
-              sawNestedFirst = true;
-            }
-          }
-          rowSet.clear();
-        }
-        assertFalse(iter.hasNext());
-        if (sawMissingPartition &&
-            sawPartitionFirst &&
-            sawPartitionLast &&
-            sawRootFirst &&
-            sawNestedFirst) {
-          // The following should appear most of the time.
-          System.out.println("All variations occurred");
-          return;
-        }
+    for (int j = 0; j < 2; j++) {
+      assertTrue(iter.hasNext());
+      rowSet = iter.next();
+
+      // Figure out which record this is and test accordingly.
+
+      RowSetReader reader = rowSet.reader();
+      assertTrue(reader.next());
+      String col1 = reader.scalar("a").getString();
+      if (col1.equals("10")) {
+        RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
+            .addRow("10", "foo", "bar", null)
+            .build();
+        RowSetUtilities.verify(expected, rowSet);
+      } else {
+        RowSet expected = new RowSetBuilder(client.allocator(), expectedSchema)
+            .addRow("20", "fred", "wilma", NESTED_DIR)
+            .build();
+        RowSetUtilities.verify(expected, rowSet);
       }
-
-      // If you see this, maybe something got fixed. Or, maybe the
-      // min parallelization hack above stopped working.
-      // Or, you were just unlucky and can try the test again.
-      // We print messages, rather than using assertTrue, to avoid
-      // introducing a flaky test.
-
-      System.out.println("Some variations did not occur");
-      System.out.println(String.format("Missing partition: %s", sawMissingPartition));
-      System.out.println(String.format("Partition first: %s", sawPartitionFirst));
-      System.out.println(String.format("Partition last: %s", sawPartitionLast));
-      System.out.println(String.format("Outer first: %s", sawRootFirst));
-      System.out.println(String.format("Nested first: %s", sawNestedFirst));
-    } finally {
-      resetV3();
-      resetMultiScan();
     }
+    assertFalse(iter.hasNext());
   }
 
   /**
@@ -286,7 +127,7 @@ public class TestPartitionRace extends BaseCsvTest {
    * "jump around" when files are shifted to a new partition depth.
    */
   @Test
-  public void testNoRaceV3() throws IOException {
+  public void testNoRace() throws IOException {
     String sql = "SELECT * FROM `dfs.data`.`%s`";
 
     TupleMetadata expectedSchema = new SchemaBuilder()
@@ -297,7 +138,6 @@ public class TestPartitionRace extends BaseCsvTest {
         .buildSchema();
 
     try {
-      enableV3(true);
       enableMultiScan();
 
       // Loop to run the query 10 times or until we see both files
@@ -363,7 +203,6 @@ public class TestPartitionRace extends BaseCsvTest {
       System.out.println(String.format("Outer first: %s", sawRootFirst));
       System.out.println(String.format("Nested first: %s", sawNestedFirst));
     } finally {
-      resetV3();
       resetMultiScan();
     }
   }
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/store/text/TestTextColumn.java b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestTextColumn.java
similarity index 91%
rename from exec/java-exec/src/test/java/org/apache/drill/exec/store/text/TestTextColumn.java
rename to exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestTextColumn.java
index 2b96924..1712a22 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/store/text/TestTextColumn.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestTextColumn.java
@@ -15,16 +15,16 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.drill.exec.store.text;
+package org.apache.drill.exec.store.easy.text.compliant;
+
+import static org.junit.Assert.assertEquals;
 
 import java.util.List;
 
-import org.apache.drill.test.BaseTestQuery;
 import org.apache.drill.exec.rpc.user.QueryDataBatch;
+import org.apache.drill.test.BaseTestQuery;
 import org.junit.Test;
 
-import static org.junit.Assert.assertEquals;
-
 public class TestTextColumn extends BaseTestQuery {
   @Test
   public void testCsvColumnSelection() throws Exception {
@@ -51,9 +51,9 @@ public class TestTextColumn extends BaseTestQuery {
         + "columns[3] as col4 from cp.`store/text/data/letters.csv`");
 
     final TestResultSet expectedResultSet = new TestResultSet();
-    expectedResultSet.addRow("a, b,", "c", "d,, \\n e", "f\\\"g");
-    expectedResultSet.addRow("d, e,", "f", "g,, \\n h", "i\\\"j");
-    expectedResultSet.addRow("g, h,", "i", "j,, \\n k", "l\\\"m");
+    expectedResultSet.addRow("a, b,", "c", "d,, \\n e", "f\"g");
+    expectedResultSet.addRow("d, e,", "f", "g,, \\n h", "i\"j");
+    expectedResultSet.addRow("g, h,", "i", "j,, \\n k", "l\"m");
 
     TestResultSet actualResultSet = new TestResultSet(actualResults);
     assertEquals(expectedResultSet, actualResultSet);
@@ -73,5 +73,4 @@ public class TestTextColumn extends BaseTestQuery {
         .sqlBaselineQuery("select COLUMNS[0], CoLuMnS[1] from cp.`store/text/data/letters.csv`")
         .go();
   }
-
 }
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/store/text/TestNewTextReader.java b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestTextReader.java
similarity index 91%
rename from exec/java-exec/src/test/java/org/apache/drill/exec/store/text/TestNewTextReader.java
rename to exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestTextReader.java
index 43ad0d8..3798206 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/store/text/TestNewTextReader.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TestTextReader.java
@@ -15,21 +15,21 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.drill.exec.store.text;
+package org.apache.drill.exec.store.easy.text.compliant;
 
 import static org.junit.Assert.assertEquals;
 import static org.junit.Assert.assertTrue;
 import static org.junit.Assert.fail;
 
-import org.apache.drill.test.BaseTestQuery;
 import org.apache.drill.categories.UnlikelyTest;
 import org.apache.drill.common.exceptions.UserRemoteException;
 import org.apache.drill.exec.proto.UserBitShared.DrillPBError.ErrorType;
+import org.apache.drill.test.BaseTestQuery;
 import org.junit.Ignore;
 import org.junit.Test;
 import org.junit.experimental.categories.Category;
 
-public class TestNewTextReader extends BaseTestQuery {
+public class TestTextReader extends BaseTestQuery {
 
   @Test
   public void fieldDelimiterWithinQuotes() throws Exception {
@@ -60,10 +60,7 @@ public class TestNewTextReader extends BaseTestQuery {
       test("select max(columns[1]) as %s from cp.`textinput/input1.csv` where %s is not null", COL_NAME, COL_NAME);
       fail("Query should have failed");
     } catch(UserRemoteException ex) {
-      assertEquals(ErrorType.DATA_READ, ex.getErrorType());
-      // Change to the following if V3 is enabled
-      // assertEquals(ErrorType.VALIDATION, ex.getErrorType());
-      assertTrue("Error message should contain " + COL_NAME, ex.getMessage().contains(COL_NAME));
+      assertEquals(ErrorType.VALIDATION, ex.getErrorType());
     }
   }
 
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/store/text/TextRecordReaderTest.java b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TextRecordReaderTest.java
similarity index 97%
rename from exec/java-exec/src/test/java/org/apache/drill/exec/store/text/TextRecordReaderTest.java
rename to exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TextRecordReaderTest.java
index fb8b50b..8e37c30 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/store/text/TextRecordReaderTest.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/store/easy/text/compliant/TextRecordReaderTest.java
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.drill.exec.store.text;
+package org.apache.drill.exec.store.easy.text.compliant;
 
 import static org.junit.Assert.assertEquals;
 
@@ -29,10 +29,9 @@ import org.apache.drill.exec.rpc.user.QueryDataBatch;
 import org.apache.drill.exec.server.Drillbit;
 import org.apache.drill.exec.server.RemoteServiceSet;
 import org.apache.drill.exec.util.VectorUtil;
-import org.junit.Test;
-
 import org.apache.drill.shaded.guava.com.google.common.base.Charsets;
 import org.apache.drill.shaded.guava.com.google.common.io.Files;
+import org.junit.Test;
 
 public class TextRecordReaderTest extends PopUnitTestBase {
 
@@ -63,5 +62,4 @@ public class TextRecordReaderTest extends PopUnitTestBase {
       assertEquals(5, count);
     }
   }
-
 }
diff --git a/exec/java-exec/src/test/resources/store/text/data/letters.csv b/exec/java-exec/src/test/resources/store/text/data/letters.csv
index 4d724e8..1ef564d 100644
--- a/exec/java-exec/src/test/resources/store/text/data/letters.csv
+++ b/exec/java-exec/src/test/resources/store/text/data/letters.csv
@@ -1,3 +1,6 @@
-"a, b,","c","d,, \n e","f\"g"
-"d, e,","f","g,, \n h","i\"j"
-"g, h,","i","j,, \n k","l\"m"
\ No newline at end of file
+# CSV uses quote as the quote escape by default.
+# Earlier versions accidentally used back-slash.
+# The back-slash, and \n, don't do anything; they are regular chars.
+"a, b,","c","d,, \n e","f""g"
+"d, e,","f","g,, \n h","i""j"
+"g, h,","i","j,, \n k","l""m"
\ No newline at end of file


[drill] 01/03: DRILL-7294: Prevent generating java beans using protostuff to avoid overriding classes with the same simple name declared as nested in the proto files

Posted by vo...@apache.org.
This is an automated email from the ASF dual-hosted git repository.

volodymyr pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/drill.git

commit 4eb726367adad2ae0981b6be311f9c4ec2f30c73
Author: Volodymyr Vysotskyi <vv...@gmail.com>
AuthorDate: Thu Jun 13 17:02:43 2019 +0300

    DRILL-7294: Prevent generating java beans using protostuff to avoid overriding classes with the same simple name declared as nested in the proto files
    
    closes #1808
---
 .../apache/drill/exec/store/jdbc/JdbcSubScan.java  |  4 +--
 .../exec/coord/store/TransientStoreConfig.java     |  2 +-
 .../impl/statistics/AvgWidthMergedStatistic.java   |  7 ++---
 .../impl/statistics/NDVMergedStatistic.java        |  7 ++---
 .../drill/exec/serialization/ProtoSerializer.java  |  4 +--
 .../exec/store/sys/PersistentStoreConfig.java      |  2 +-
 exec/jdbc-all/pom.xml                              |  4 +--
 protocol/pom.xml                                   | 35 +++++-----------------
 8 files changed, 21 insertions(+), 44 deletions(-)

diff --git a/contrib/storage-jdbc/src/main/java/org/apache/drill/exec/store/jdbc/JdbcSubScan.java b/contrib/storage-jdbc/src/main/java/org/apache/drill/exec/store/jdbc/JdbcSubScan.java
index c9d5f0d..43bf909 100755
--- a/contrib/storage-jdbc/src/main/java/org/apache/drill/exec/store/jdbc/JdbcSubScan.java
+++ b/contrib/storage-jdbc/src/main/java/org/apache/drill/exec/store/jdbc/JdbcSubScan.java
@@ -21,7 +21,7 @@ import org.apache.drill.common.exceptions.ExecutionSetupException;
 import org.apache.drill.common.expression.SchemaPath;
 import org.apache.drill.common.logical.StoragePluginConfig;
 import org.apache.drill.exec.physical.base.AbstractSubScan;
-import org.apache.drill.exec.proto.beans.CoreOperatorType;
+import org.apache.drill.exec.proto.UserBitShared;
 import org.apache.drill.exec.store.StoragePluginRegistry;
 
 import com.fasterxml.jackson.annotation.JacksonInject;
@@ -60,7 +60,7 @@ public class JdbcSubScan extends AbstractSubScan {
 
   @Override
   public int getOperatorType() {
-    return CoreOperatorType.JDBC_SCAN.getNumber();
+    return UserBitShared.CoreOperatorType.JDBC_SCAN.getNumber();
   }
 
   public String getSql() {
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/coord/store/TransientStoreConfig.java b/exec/java-exec/src/main/java/org/apache/drill/exec/coord/store/TransientStoreConfig.java
index 3c95639..9622a77 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/coord/store/TransientStoreConfig.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/coord/store/TransientStoreConfig.java
@@ -17,8 +17,8 @@
  */
 package org.apache.drill.exec.coord.store;
 
-import com.dyuproject.protostuff.Schema;
 import com.fasterxml.jackson.databind.ObjectMapper;
+import io.protostuff.Schema;
 import org.apache.drill.shaded.guava.com.google.common.base.Objects;
 import org.apache.drill.shaded.guava.com.google.common.base.Preconditions;
 import org.apache.drill.shaded.guava.com.google.common.base.Strings;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/statistics/AvgWidthMergedStatistic.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/statistics/AvgWidthMergedStatistic.java
index 6a9660d..0a01e74 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/statistics/AvgWidthMergedStatistic.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/statistics/AvgWidthMergedStatistic.java
@@ -23,7 +23,6 @@ import java.io.IOException;
 import java.util.HashMap;
 import java.util.List;
 import java.util.Map;
-import org.apache.drill.common.types.MinorType;
 import org.apache.drill.common.types.TypeProtos;
 import org.apache.drill.exec.record.MajorTypeSerDe;
 import org.apache.drill.exec.vector.NullableFloat8Vector;
@@ -139,9 +138,9 @@ public class AvgWidthMergedStatistic extends AbstractMergedStatistic {
     }
     // If variable length type - then use the nonNullCount. Otherwise, use the Count,
     // since even NULL values take up the same space.
-    if (type == MinorType.VAR16CHAR.getNumber()
-        || type == MinorType.VARCHAR.getNumber()
-        || type == MinorType.VARBINARY.getNumber()) {
+    if (type == TypeProtos.MinorType.VAR16CHAR.getNumber()
+        || type == TypeProtos.MinorType.VARCHAR.getNumber()
+        || type == TypeProtos.MinorType.VARBINARY.getNumber()) {
       return nonNullStatCounts.getStat(colName);
     } else {
       return statCounts.getStat(colName);
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/statistics/NDVMergedStatistic.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/statistics/NDVMergedStatistic.java
index fc87fbd..88e93d4 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/statistics/NDVMergedStatistic.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/statistics/NDVMergedStatistic.java
@@ -29,7 +29,6 @@ import java.io.IOException;
 import java.util.HashMap;
 import java.util.List;
 import java.util.Map;
-import org.apache.drill.common.types.MinorType;
 import org.apache.drill.common.types.TypeProtos;
 import org.apache.drill.exec.record.MajorTypeSerDe;
 import org.apache.drill.exec.server.options.OptionManager;
@@ -187,9 +186,9 @@ public class NDVMergedStatistic extends AbstractMergedStatistic {
     }
     // If variable length type - then use the nonNullCount. Otherwise, use the Count,
     // since even NULL values take up the same space.
-    if (type == MinorType.VAR16CHAR.getNumber()
-            || type == MinorType.VARCHAR.getNumber()
-            || type == MinorType.VARBINARY.getNumber()) {
+    if (type == TypeProtos.MinorType.VAR16CHAR.getNumber()
+            || type == TypeProtos.MinorType.VARCHAR.getNumber()
+            || type == TypeProtos.MinorType.VARBINARY.getNumber()) {
       return nonNullStatCounts.getStat(colName);
     } else {
       return statCounts.getStat(colName);
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/serialization/ProtoSerializer.java b/exec/java-exec/src/main/java/org/apache/drill/exec/serialization/ProtoSerializer.java
index 51878d0..0124292 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/serialization/ProtoSerializer.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/serialization/ProtoSerializer.java
@@ -20,8 +20,8 @@ package org.apache.drill.exec.serialization;
 import java.io.ByteArrayOutputStream;
 import java.io.IOException;
 
-import com.dyuproject.protostuff.JsonIOUtil;
-import com.dyuproject.protostuff.Schema;
+import io.protostuff.JsonIOUtil;
+import io.protostuff.Schema;
 import org.apache.drill.shaded.guava.com.google.common.base.Objects;
 import org.apache.drill.shaded.guava.com.google.common.base.Preconditions;
 import com.google.protobuf.Message;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/sys/PersistentStoreConfig.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/sys/PersistentStoreConfig.java
index 532033a..668306e 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/sys/PersistentStoreConfig.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/sys/PersistentStoreConfig.java
@@ -17,12 +17,12 @@
  */
 package org.apache.drill.exec.store.sys;
 
-import com.dyuproject.protostuff.Schema;
 import com.fasterxml.jackson.databind.ObjectMapper;
 import org.apache.drill.shaded.guava.com.google.common.base.Objects;
 import org.apache.drill.shaded.guava.com.google.common.base.Preconditions;
 import com.google.protobuf.Message;
 import com.google.protobuf.Message.Builder;
+import io.protostuff.Schema;
 import org.apache.drill.exec.serialization.InstanceSerializer;
 import org.apache.drill.exec.serialization.JacksonSerializer;
 import org.apache.drill.exec.serialization.ProtoSerializer;
diff --git a/exec/jdbc-all/pom.xml b/exec/jdbc-all/pom.xml
index be79b41..a7effe3 100644
--- a/exec/jdbc-all/pom.xml
+++ b/exec/jdbc-all/pom.xml
@@ -299,7 +299,7 @@
               <include>*:*</include>
             </includes>
             <excludes>
-              <exclude>com.dyuproject.protostuff:*</exclude>
+              <exclude>io.protostuff:*</exclude>
               <exclude>org.apache.calcite:calcite-core</exclude>
               <exclude>org.apache.calcite:calcite-linq4j</exclude>
               <exclude>org.pentaho:*</exclude>
@@ -614,7 +614,7 @@
                   </includes>
                   <excludes>
                     <exclude>org.slf4j:jcl-over-slf4j</exclude>
-                    <exclude>com.dyuproject.protostuff:*</exclude>
+                    <exclude>io.protostuff:*</exclude>
                     <exclude>org.apache.calcite:calcite-core</exclude>
                     <exclude>org.apache.calcite:calcite-linq4j</exclude>
                     <exclude>org.pentaho:*</exclude>
diff --git a/protocol/pom.xml b/protocol/pom.xml
index 04629ac..73226a3 100644
--- a/protocol/pom.xml
+++ b/protocol/pom.xml
@@ -37,14 +37,14 @@
       <artifactId>protobuf-java</artifactId>
     </dependency>
     <dependency>
-      <groupId>com.dyuproject.protostuff</groupId>
+      <groupId>io.protostuff</groupId>
       <artifactId>protostuff-core</artifactId>
-      <version>1.0.8</version>
+      <version>1.6.0</version>
     </dependency>
     <dependency>
-      <groupId>com.dyuproject.protostuff</groupId>
+      <groupId>io.protostuff</groupId>
       <artifactId>protostuff-json</artifactId>
-      <version>1.0.8</version>
+      <version>1.6.0</version>
     </dependency>
   </dependencies>
 
@@ -82,9 +82,9 @@
             </executions>
           </plugin>
           <plugin>
-            <groupId>com.dyuproject.protostuff</groupId>
+            <groupId>io.protostuff</groupId>
             <artifactId>protostuff-maven-plugin</artifactId>
-            <version>1.0.8</version>
+            <version>1.6.0</version>
             <configuration>
               <protoModules>
                 <protoModule>
@@ -98,27 +98,6 @@
                     </property>
                   </options>
                 </protoModule>
-                <protoModule>
-                  <source>src/main/protobuf/</source>
-                  <outputDir>src/main/java</outputDir>
-                  <output>java_bean</output>
-                  <encoding>UTF-8</encoding>
-                  <options>
-                    <property>
-                      <name>generate_field_map</name>
-                    </property>
-                    <property>
-                      <name>org.apache.drill.exec.proto</name>
-                      <value>org.apache.drill.exec.proto.beans</value>
-                    </property>
-                    <property>
-                      <name>primitive_numbers_if_optional</name>
-                    </property>
-                    <property>
-                      <name>builder_pattern</name>
-                    </property>
-                  </options>
-                </protoModule>
               </protoModules>
             </configuration>
             <executions>
@@ -134,7 +113,7 @@
           <plugin>
             <groupId>com.github.igor-petruk.protobuf</groupId>
             <artifactId>protobuf-maven-plugin</artifactId>
-            <version>0.6.3</version>
+            <version>0.6.5</version>
             <executions>
               <execution>
                 <goals>