You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@accumulo.apache.org by mm...@apache.org on 2016/12/13 18:22:45 UTC

[1/2] accumulo-wikisearch git commit: ACCUMULO-2446: many updates for Accumulo 1.8.0 [Forced Update!]

Repository: accumulo-wikisearch
Updated Branches:
  refs/heads/1.8 dd23f1071 -> e32632caf (forced update)


http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/e32632ca/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/protobuf/Uid.java
----------------------------------------------------------------------
diff --git a/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/protobuf/Uid.java b/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/protobuf/Uid.java
index c469748..619be73 100644
--- a/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/protobuf/Uid.java
+++ b/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/protobuf/Uid.java
@@ -1,19 +1,3 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one or more
- * contributor license agreements.  See the NOTICE file distributed with
- * this work for additional information regarding copyright ownership.
- * The ASF licenses this file to You under the Apache License, Version 2.0
- * (the "License"); you may not use this file except in compliance with
- * the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
 // Generated by the protocol buffer compiler.  DO NOT EDIT!
 // source: Uid.proto
 
@@ -21,119 +5,282 @@ package org.apache.accumulo.examples.wikisearch.protobuf;
 
 public final class Uid {
   private Uid() {}
-  
-  public static void registerAllExtensions(com.google.protobuf.ExtensionRegistry registry) {}
-  
-  public static final class List extends com.google.protobuf.GeneratedMessage {
+  public static void registerAllExtensions(
+      com.google.protobuf.ExtensionRegistry registry) {
+  }
+  public interface ListOrBuilder
+      extends com.google.protobuf.MessageOrBuilder {
+
+    // required bool IGNORE = 1;
+    /**
+     * <code>required bool IGNORE = 1;</code>
+     */
+    boolean hasIGNORE();
+    /**
+     * <code>required bool IGNORE = 1;</code>
+     */
+    boolean getIGNORE();
+
+    // required uint64 COUNT = 2;
+    /**
+     * <code>required uint64 COUNT = 2;</code>
+     */
+    boolean hasCOUNT();
+    /**
+     * <code>required uint64 COUNT = 2;</code>
+     */
+    long getCOUNT();
+
+    // repeated string UID = 3;
+    /**
+     * <code>repeated string UID = 3;</code>
+     */
+    java.util.List<java.lang.String>
+    getUIDList();
+    /**
+     * <code>repeated string UID = 3;</code>
+     */
+    int getUIDCount();
+    /**
+     * <code>repeated string UID = 3;</code>
+     */
+    java.lang.String getUID(int index);
+    /**
+     * <code>repeated string UID = 3;</code>
+     */
+    com.google.protobuf.ByteString
+        getUIDBytes(int index);
+  }
+  /**
+   * Protobuf type {@code org.apache.accumulo.examples.wikisearch.protobuf.List}
+   */
+  public static final class List extends
+      com.google.protobuf.GeneratedMessage
+      implements ListOrBuilder {
     // Use List.newBuilder() to construct.
-    private List() {
-      initFields();
+    private List(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+      super(builder);
+      this.unknownFields = builder.getUnknownFields();
     }
-    
-    private List(boolean noInit) {}
-    
+    private List(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
     private static final List defaultInstance;
-    
     public static List getDefaultInstance() {
       return defaultInstance;
     }
-    
+
     public List getDefaultInstanceForType() {
       return defaultInstance;
     }
-    
-    public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() {
-      return org.apache.accumulo.examples.wikisearch.protobuf.Uid.internal_static_protobuf_List_descriptor;
+
+    private final com.google.protobuf.UnknownFieldSet unknownFields;
+    @java.lang.Override
+    public final com.google.protobuf.UnknownFieldSet
+        getUnknownFields() {
+      return this.unknownFields;
     }
-    
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() {
-      return org.apache.accumulo.examples.wikisearch.protobuf.Uid.internal_static_protobuf_List_fieldAccessorTable;
+    private List(
+        com.google.protobuf.CodedInputStream input,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      initFields();
+      int mutable_bitField0_ = 0;
+      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+          com.google.protobuf.UnknownFieldSet.newBuilder();
+      try {
+        boolean done = false;
+        while (!done) {
+          int tag = input.readTag();
+          switch (tag) {
+            case 0:
+              done = true;
+              break;
+            default: {
+              if (!parseUnknownField(input, unknownFields,
+                                     extensionRegistry, tag)) {
+                done = true;
+              }
+              break;
+            }
+            case 8: {
+              bitField0_ |= 0x00000001;
+              iGNORE_ = input.readBool();
+              break;
+            }
+            case 16: {
+              bitField0_ |= 0x00000002;
+              cOUNT_ = input.readUInt64();
+              break;
+            }
+            case 26: {
+              if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
+                uID_ = new com.google.protobuf.LazyStringArrayList();
+                mutable_bitField0_ |= 0x00000004;
+              }
+              uID_.add(input.readBytes());
+              break;
+            }
+          }
+        }
+      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+        throw e.setUnfinishedMessage(this);
+      } catch (java.io.IOException e) {
+        throw new com.google.protobuf.InvalidProtocolBufferException(
+            e.getMessage()).setUnfinishedMessage(this);
+      } finally {
+        if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
+          uID_ = new com.google.protobuf.UnmodifiableLazyStringList(uID_);
+        }
+        this.unknownFields = unknownFields.build();
+        makeExtensionsImmutable();
+      }
     }
-    
+    public static final com.google.protobuf.Descriptors.Descriptor
+        getDescriptor() {
+      return org.apache.accumulo.examples.wikisearch.protobuf.Uid.internal_static_org_apache_accumulo_examples_wikisearch_protobuf_List_descriptor;
+    }
+
+    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+        internalGetFieldAccessorTable() {
+      return org.apache.accumulo.examples.wikisearch.protobuf.Uid.internal_static_org_apache_accumulo_examples_wikisearch_protobuf_List_fieldAccessorTable
+          .ensureFieldAccessorsInitialized(
+              org.apache.accumulo.examples.wikisearch.protobuf.Uid.List.class, org.apache.accumulo.examples.wikisearch.protobuf.Uid.List.Builder.class);
+    }
+
+    public static com.google.protobuf.Parser<List> PARSER =
+        new com.google.protobuf.AbstractParser<List>() {
+      public List parsePartialFrom(
+          com.google.protobuf.CodedInputStream input,
+          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+          throws com.google.protobuf.InvalidProtocolBufferException {
+        return new List(input, extensionRegistry);
+      }
+    };
+
+    @java.lang.Override
+    public com.google.protobuf.Parser<List> getParserForType() {
+      return PARSER;
+    }
+
+    private int bitField0_;
     // required bool IGNORE = 1;
     public static final int IGNORE_FIELD_NUMBER = 1;
-    private boolean hasIGNORE;
-    private boolean iGNORE_ = false;
-    
+    private boolean iGNORE_;
+    /**
+     * <code>required bool IGNORE = 1;</code>
+     */
     public boolean hasIGNORE() {
-      return hasIGNORE;
+      return ((bitField0_ & 0x00000001) == 0x00000001);
     }
-    
+    /**
+     * <code>required bool IGNORE = 1;</code>
+     */
     public boolean getIGNORE() {
       return iGNORE_;
     }
-    
+
     // required uint64 COUNT = 2;
     public static final int COUNT_FIELD_NUMBER = 2;
-    private boolean hasCOUNT;
-    private long cOUNT_ = 0L;
-    
+    private long cOUNT_;
+    /**
+     * <code>required uint64 COUNT = 2;</code>
+     */
     public boolean hasCOUNT() {
-      return hasCOUNT;
+      return ((bitField0_ & 0x00000002) == 0x00000002);
     }
-    
+    /**
+     * <code>required uint64 COUNT = 2;</code>
+     */
     public long getCOUNT() {
       return cOUNT_;
     }
-    
+
     // repeated string UID = 3;
     public static final int UID_FIELD_NUMBER = 3;
-    private java.util.List<java.lang.String> uID_ = java.util.Collections.emptyList();
-    
-    public java.util.List<java.lang.String> getUIDList() {
+    private com.google.protobuf.LazyStringList uID_;
+    /**
+     * <code>repeated string UID = 3;</code>
+     */
+    public java.util.List<java.lang.String>
+        getUIDList() {
       return uID_;
     }
-    
+    /**
+     * <code>repeated string UID = 3;</code>
+     */
     public int getUIDCount() {
       return uID_.size();
     }
-    
+    /**
+     * <code>repeated string UID = 3;</code>
+     */
     public java.lang.String getUID(int index) {
       return uID_.get(index);
     }
-    
-    private void initFields() {}
-    
+    /**
+     * <code>repeated string UID = 3;</code>
+     */
+    public com.google.protobuf.ByteString
+        getUIDBytes(int index) {
+      return uID_.getByteString(index);
+    }
+
+    private void initFields() {
+      iGNORE_ = false;
+      cOUNT_ = 0L;
+      uID_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+    }
+    private byte memoizedIsInitialized = -1;
     public final boolean isInitialized() {
-      if (!hasIGNORE)
+      byte isInitialized = memoizedIsInitialized;
+      if (isInitialized != -1) return isInitialized == 1;
+
+      if (!hasIGNORE()) {
+        memoizedIsInitialized = 0;
         return false;
-      if (!hasCOUNT)
+      }
+      if (!hasCOUNT()) {
+        memoizedIsInitialized = 0;
         return false;
+      }
+      memoizedIsInitialized = 1;
       return true;
     }
-    
-    public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException {
+
+    public void writeTo(com.google.protobuf.CodedOutputStream output)
+                        throws java.io.IOException {
       getSerializedSize();
-      if (hasIGNORE()) {
-        output.writeBool(1, getIGNORE());
+      if (((bitField0_ & 0x00000001) == 0x00000001)) {
+        output.writeBool(1, iGNORE_);
       }
-      if (hasCOUNT()) {
-        output.writeUInt64(2, getCOUNT());
+      if (((bitField0_ & 0x00000002) == 0x00000002)) {
+        output.writeUInt64(2, cOUNT_);
       }
-      for (java.lang.String element : getUIDList()) {
-        output.writeString(3, element);
+      for (int i = 0; i < uID_.size(); i++) {
+        output.writeBytes(3, uID_.getByteString(i));
       }
       getUnknownFields().writeTo(output);
     }
-    
+
     private int memoizedSerializedSize = -1;
-    
     public int getSerializedSize() {
       int size = memoizedSerializedSize;
-      if (size != -1)
-        return size;
-      
+      if (size != -1) return size;
+
       size = 0;
-      if (hasIGNORE()) {
-        size += com.google.protobuf.CodedOutputStream.computeBoolSize(1, getIGNORE());
+      if (((bitField0_ & 0x00000001) == 0x00000001)) {
+        size += com.google.protobuf.CodedOutputStream
+          .computeBoolSize(1, iGNORE_);
       }
-      if (hasCOUNT()) {
-        size += com.google.protobuf.CodedOutputStream.computeUInt64Size(2, getCOUNT());
+      if (((bitField0_ & 0x00000002) == 0x00000002)) {
+        size += com.google.protobuf.CodedOutputStream
+          .computeUInt64Size(2, cOUNT_);
       }
       {
         int dataSize = 0;
-        for (java.lang.String element : getUIDList()) {
-          dataSize += com.google.protobuf.CodedOutputStream.computeStringSizeNoTag(element);
+        for (int i = 0; i < uID_.size(); i++) {
+          dataSize += com.google.protobuf.CodedOutputStream
+            .computeBytesSizeNoTag(uID_.getByteString(i));
         }
         size += dataSize;
         size += 1 * getUIDList().size();
@@ -142,156 +289,182 @@ public final class Uid {
       memoizedSerializedSize = size;
       return size;
     }
-    
-    public static org.apache.accumulo.examples.wikisearch.protobuf.Uid.List parseFrom(com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data).buildParsed();
+
+    private static final long serialVersionUID = 0L;
+    @java.lang.Override
+    protected java.lang.Object writeReplace()
+        throws java.io.ObjectStreamException {
+      return super.writeReplace();
     }
-    
-    public static org.apache.accumulo.examples.wikisearch.protobuf.Uid.List parseFrom(com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+
+    public static org.apache.accumulo.examples.wikisearch.protobuf.Uid.List parseFrom(
+        com.google.protobuf.ByteString data)
         throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data, extensionRegistry).buildParsed();
+      return PARSER.parseFrom(data);
     }
-    
-    public static org.apache.accumulo.examples.wikisearch.protobuf.Uid.List parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data).buildParsed();
+    public static org.apache.accumulo.examples.wikisearch.protobuf.Uid.List parseFrom(
+        com.google.protobuf.ByteString data,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data, extensionRegistry);
     }
-    
-    public static org.apache.accumulo.examples.wikisearch.protobuf.Uid.List parseFrom(byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+    public static org.apache.accumulo.examples.wikisearch.protobuf.Uid.List parseFrom(byte[] data)
         throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data, extensionRegistry).buildParsed();
+      return PARSER.parseFrom(data);
     }
-    
-    public static org.apache.accumulo.examples.wikisearch.protobuf.Uid.List parseFrom(java.io.InputStream input) throws java.io.IOException {
-      return newBuilder().mergeFrom(input).buildParsed();
+    public static org.apache.accumulo.examples.wikisearch.protobuf.Uid.List parseFrom(
+        byte[] data,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data, extensionRegistry);
     }
-    
-    public static org.apache.accumulo.examples.wikisearch.protobuf.Uid.List parseFrom(java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+    public static org.apache.accumulo.examples.wikisearch.protobuf.Uid.List parseFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return newBuilder().mergeFrom(input, extensionRegistry).buildParsed();
-    }
-    
-    public static org.apache.accumulo.examples.wikisearch.protobuf.Uid.List parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException {
-      Builder builder = newBuilder();
-      if (builder.mergeDelimitedFrom(input)) {
-        return builder.buildParsed();
-      } else {
-        return null;
-      }
+      return PARSER.parseFrom(input);
     }
-    
-    public static org.apache.accumulo.examples.wikisearch.protobuf.Uid.List parseDelimitedFrom(java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+    public static org.apache.accumulo.examples.wikisearch.protobuf.Uid.List parseFrom(
+        java.io.InputStream input,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      Builder builder = newBuilder();
-      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
-        return builder.buildParsed();
-      } else {
-        return null;
-      }
+      return PARSER.parseFrom(input, extensionRegistry);
     }
-    
-    public static org.apache.accumulo.examples.wikisearch.protobuf.Uid.List parseFrom(com.google.protobuf.CodedInputStream input) throws java.io.IOException {
-      return newBuilder().mergeFrom(input).buildParsed();
+    public static org.apache.accumulo.examples.wikisearch.protobuf.Uid.List parseDelimitedFrom(java.io.InputStream input)
+        throws java.io.IOException {
+      return PARSER.parseDelimitedFrom(input);
     }
-    
-    public static org.apache.accumulo.examples.wikisearch.protobuf.Uid.List parseFrom(com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+    public static org.apache.accumulo.examples.wikisearch.protobuf.Uid.List parseDelimitedFrom(
+        java.io.InputStream input,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return newBuilder().mergeFrom(input, extensionRegistry).buildParsed();
+      return PARSER.parseDelimitedFrom(input, extensionRegistry);
     }
-    
-    public static Builder newBuilder() {
-      return Builder.create();
+    public static org.apache.accumulo.examples.wikisearch.protobuf.Uid.List parseFrom(
+        com.google.protobuf.CodedInputStream input)
+        throws java.io.IOException {
+      return PARSER.parseFrom(input);
     }
-    
-    public Builder newBuilderForType() {
-      return newBuilder();
+    public static org.apache.accumulo.examples.wikisearch.protobuf.Uid.List parseFrom(
+        com.google.protobuf.CodedInputStream input,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws java.io.IOException {
+      return PARSER.parseFrom(input, extensionRegistry);
     }
-    
+
+    public static Builder newBuilder() { return Builder.create(); }
+    public Builder newBuilderForType() { return newBuilder(); }
     public static Builder newBuilder(org.apache.accumulo.examples.wikisearch.protobuf.Uid.List prototype) {
       return newBuilder().mergeFrom(prototype);
     }
-    
-    public Builder toBuilder() {
-      return newBuilder(this);
-    }
-    
-    public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> {
-      private org.apache.accumulo.examples.wikisearch.protobuf.Uid.List result;
-      
-      // Construct using protobuf.Uid.List.newBuilder()
-      private Builder() {}
-      
-      private static Builder create() {
-        Builder builder = new Builder();
-        builder.result = new org.apache.accumulo.examples.wikisearch.protobuf.Uid.List();
-        return builder;
+    public Builder toBuilder() { return newBuilder(this); }
+
+    @java.lang.Override
+    protected Builder newBuilderForType(
+        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+      Builder builder = new Builder(parent);
+      return builder;
+    }
+    /**
+     * Protobuf type {@code org.apache.accumulo.examples.wikisearch.protobuf.List}
+     */
+    public static final class Builder extends
+        com.google.protobuf.GeneratedMessage.Builder<Builder>
+       implements org.apache.accumulo.examples.wikisearch.protobuf.Uid.ListOrBuilder {
+      public static final com.google.protobuf.Descriptors.Descriptor
+          getDescriptor() {
+        return org.apache.accumulo.examples.wikisearch.protobuf.Uid.internal_static_org_apache_accumulo_examples_wikisearch_protobuf_List_descriptor;
       }
-      
-      protected org.apache.accumulo.examples.wikisearch.protobuf.Uid.List internalGetResult() {
-        return result;
+
+      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+          internalGetFieldAccessorTable() {
+        return org.apache.accumulo.examples.wikisearch.protobuf.Uid.internal_static_org_apache_accumulo_examples_wikisearch_protobuf_List_fieldAccessorTable
+            .ensureFieldAccessorsInitialized(
+                org.apache.accumulo.examples.wikisearch.protobuf.Uid.List.class, org.apache.accumulo.examples.wikisearch.protobuf.Uid.List.Builder.class);
       }
-      
-      public Builder clear() {
-        if (result == null) {
-          throw new IllegalStateException("Cannot call clear() after build().");
+
+      // Construct using org.apache.accumulo.examples.wikisearch.protobuf.Uid.List.newBuilder()
+      private Builder() {
+        maybeForceBuilderInitialization();
+      }
+
+      private Builder(
+          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+        super(parent);
+        maybeForceBuilderInitialization();
+      }
+      private void maybeForceBuilderInitialization() {
+        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
         }
-        result = new org.apache.accumulo.examples.wikisearch.protobuf.Uid.List();
+      }
+      private static Builder create() {
+        return new Builder();
+      }
+
+      public Builder clear() {
+        super.clear();
+        iGNORE_ = false;
+        bitField0_ = (bitField0_ & ~0x00000001);
+        cOUNT_ = 0L;
+        bitField0_ = (bitField0_ & ~0x00000002);
+        uID_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+        bitField0_ = (bitField0_ & ~0x00000004);
         return this;
       }
-      
+
       public Builder clone() {
-        return create().mergeFrom(result);
+        return create().mergeFrom(buildPartial());
       }
-      
-      public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() {
-        return org.apache.accumulo.examples.wikisearch.protobuf.Uid.List.getDescriptor();
+
+      public com.google.protobuf.Descriptors.Descriptor
+          getDescriptorForType() {
+        return org.apache.accumulo.examples.wikisearch.protobuf.Uid.internal_static_org_apache_accumulo_examples_wikisearch_protobuf_List_descriptor;
       }
-      
+
       public org.apache.accumulo.examples.wikisearch.protobuf.Uid.List getDefaultInstanceForType() {
         return org.apache.accumulo.examples.wikisearch.protobuf.Uid.List.getDefaultInstance();
       }
-      
-      public boolean isInitialized() {
-        return result.isInitialized();
-      }
-      
+
       public org.apache.accumulo.examples.wikisearch.protobuf.Uid.List build() {
-        if (result != null && !isInitialized()) {
+        org.apache.accumulo.examples.wikisearch.protobuf.Uid.List result = buildPartial();
+        if (!result.isInitialized()) {
           throw newUninitializedMessageException(result);
         }
-        return buildPartial();
-      }
-      
-      private org.apache.accumulo.examples.wikisearch.protobuf.Uid.List buildParsed() throws com.google.protobuf.InvalidProtocolBufferException {
-        if (!isInitialized()) {
-          throw newUninitializedMessageException(result).asInvalidProtocolBufferException();
-        }
-        return buildPartial();
+        return result;
       }
-      
+
       public org.apache.accumulo.examples.wikisearch.protobuf.Uid.List buildPartial() {
-        if (result == null) {
-          throw new IllegalStateException("build() has already been called on this Builder.");
+        org.apache.accumulo.examples.wikisearch.protobuf.Uid.List result = new org.apache.accumulo.examples.wikisearch.protobuf.Uid.List(this);
+        int from_bitField0_ = bitField0_;
+        int to_bitField0_ = 0;
+        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+          to_bitField0_ |= 0x00000001;
+        }
+        result.iGNORE_ = iGNORE_;
+        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
+          to_bitField0_ |= 0x00000002;
         }
-        if (result.uID_ != java.util.Collections.EMPTY_LIST) {
-          result.uID_ = java.util.Collections.unmodifiableList(result.uID_);
+        result.cOUNT_ = cOUNT_;
+        if (((bitField0_ & 0x00000004) == 0x00000004)) {
+          uID_ = new com.google.protobuf.UnmodifiableLazyStringList(
+              uID_);
+          bitField0_ = (bitField0_ & ~0x00000004);
         }
-        org.apache.accumulo.examples.wikisearch.protobuf.Uid.List returnMe = result;
-        result = null;
-        return returnMe;
+        result.uID_ = uID_;
+        result.bitField0_ = to_bitField0_;
+        onBuilt();
+        return result;
       }
-      
+
       public Builder mergeFrom(com.google.protobuf.Message other) {
         if (other instanceof org.apache.accumulo.examples.wikisearch.protobuf.Uid.List) {
-          return mergeFrom((org.apache.accumulo.examples.wikisearch.protobuf.Uid.List) other);
+          return mergeFrom((org.apache.accumulo.examples.wikisearch.protobuf.Uid.List)other);
         } else {
           super.mergeFrom(other);
           return this;
         }
       }
-      
+
       public Builder mergeFrom(org.apache.accumulo.examples.wikisearch.protobuf.Uid.List other) {
-        if (other == org.apache.accumulo.examples.wikisearch.protobuf.Uid.List.getDefaultInstance())
-          return this;
+        if (other == org.apache.accumulo.examples.wikisearch.protobuf.Uid.List.getDefaultInstance()) return this;
         if (other.hasIGNORE()) {
           setIGNORE(other.getIGNORE());
         }
@@ -299,172 +472,259 @@ public final class Uid {
           setCOUNT(other.getCOUNT());
         }
         if (!other.uID_.isEmpty()) {
-          if (result.uID_.isEmpty()) {
-            result.uID_ = new java.util.ArrayList<java.lang.String>();
+          if (uID_.isEmpty()) {
+            uID_ = other.uID_;
+            bitField0_ = (bitField0_ & ~0x00000004);
+          } else {
+            ensureUIDIsMutable();
+            uID_.addAll(other.uID_);
           }
-          result.uID_.addAll(other.uID_);
+          onChanged();
         }
         this.mergeUnknownFields(other.getUnknownFields());
         return this;
       }
-      
-      public Builder mergeFrom(com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+
+      public final boolean isInitialized() {
+        if (!hasIGNORE()) {
+          
+          return false;
+        }
+        if (!hasCOUNT()) {
+          
+          return false;
+        }
+        return true;
+      }
+
+      public Builder mergeFrom(
+          com.google.protobuf.CodedInputStream input,
+          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
           throws java.io.IOException {
-        com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(this.getUnknownFields());
-        while (true) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              this.setUnknownFields(unknownFields.build());
-              return this;
-            default: {
-              if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) {
-                this.setUnknownFields(unknownFields.build());
-                return this;
-              }
-              break;
-            }
-            case 8: {
-              setIGNORE(input.readBool());
-              break;
-            }
-            case 16: {
-              setCOUNT(input.readUInt64());
-              break;
-            }
-            case 26: {
-              addUID(input.readString());
-              break;
-            }
+        org.apache.accumulo.examples.wikisearch.protobuf.Uid.List parsedMessage = null;
+        try {
+          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+          parsedMessage = (org.apache.accumulo.examples.wikisearch.protobuf.Uid.List) e.getUnfinishedMessage();
+          throw e;
+        } finally {
+          if (parsedMessage != null) {
+            mergeFrom(parsedMessage);
           }
         }
+        return this;
       }
-      
+      private int bitField0_;
+
       // required bool IGNORE = 1;
+      private boolean iGNORE_ ;
+      /**
+       * <code>required bool IGNORE = 1;</code>
+       */
       public boolean hasIGNORE() {
-        return result.hasIGNORE();
+        return ((bitField0_ & 0x00000001) == 0x00000001);
       }
-      
+      /**
+       * <code>required bool IGNORE = 1;</code>
+       */
       public boolean getIGNORE() {
-        return result.getIGNORE();
+        return iGNORE_;
       }
-      
+      /**
+       * <code>required bool IGNORE = 1;</code>
+       */
       public Builder setIGNORE(boolean value) {
-        result.hasIGNORE = true;
-        result.iGNORE_ = value;
+        bitField0_ |= 0x00000001;
+        iGNORE_ = value;
+        onChanged();
         return this;
       }
-      
+      /**
+       * <code>required bool IGNORE = 1;</code>
+       */
       public Builder clearIGNORE() {
-        result.hasIGNORE = false;
-        result.iGNORE_ = false;
+        bitField0_ = (bitField0_ & ~0x00000001);
+        iGNORE_ = false;
+        onChanged();
         return this;
       }
-      
+
       // required uint64 COUNT = 2;
+      private long cOUNT_ ;
+      /**
+       * <code>required uint64 COUNT = 2;</code>
+       */
       public boolean hasCOUNT() {
-        return result.hasCOUNT();
+        return ((bitField0_ & 0x00000002) == 0x00000002);
       }
-      
+      /**
+       * <code>required uint64 COUNT = 2;</code>
+       */
       public long getCOUNT() {
-        return result.getCOUNT();
+        return cOUNT_;
       }
-      
+      /**
+       * <code>required uint64 COUNT = 2;</code>
+       */
       public Builder setCOUNT(long value) {
-        result.hasCOUNT = true;
-        result.cOUNT_ = value;
+        bitField0_ |= 0x00000002;
+        cOUNT_ = value;
+        onChanged();
         return this;
       }
-      
+      /**
+       * <code>required uint64 COUNT = 2;</code>
+       */
       public Builder clearCOUNT() {
-        result.hasCOUNT = false;
-        result.cOUNT_ = 0L;
+        bitField0_ = (bitField0_ & ~0x00000002);
+        cOUNT_ = 0L;
+        onChanged();
         return this;
       }
-      
+
       // repeated string UID = 3;
-      public java.util.List<java.lang.String> getUIDList() {
-        return java.util.Collections.unmodifiableList(result.uID_);
-      }
-      
+      private com.google.protobuf.LazyStringList uID_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+      private void ensureUIDIsMutable() {
+        if (!((bitField0_ & 0x00000004) == 0x00000004)) {
+          uID_ = new com.google.protobuf.LazyStringArrayList(uID_);
+          bitField0_ |= 0x00000004;
+         }
+      }
+      /**
+       * <code>repeated string UID = 3;</code>
+       */
+      public java.util.List<java.lang.String>
+          getUIDList() {
+        return java.util.Collections.unmodifiableList(uID_);
+      }
+      /**
+       * <code>repeated string UID = 3;</code>
+       */
       public int getUIDCount() {
-        return result.getUIDCount();
+        return uID_.size();
       }
-      
+      /**
+       * <code>repeated string UID = 3;</code>
+       */
       public java.lang.String getUID(int index) {
-        return result.getUID(index);
-      }
-      
-      public Builder setUID(int index, java.lang.String value) {
+        return uID_.get(index);
+      }
+      /**
+       * <code>repeated string UID = 3;</code>
+       */
+      public com.google.protobuf.ByteString
+          getUIDBytes(int index) {
+        return uID_.getByteString(index);
+      }
+      /**
+       * <code>repeated string UID = 3;</code>
+       */
+      public Builder setUID(
+          int index, java.lang.String value) {
         if (value == null) {
-          throw new NullPointerException();
-        }
-        result.uID_.set(index, value);
+    throw new NullPointerException();
+  }
+  ensureUIDIsMutable();
+        uID_.set(index, value);
+        onChanged();
         return this;
       }
-      
-      public Builder addUID(java.lang.String value) {
+      /**
+       * <code>repeated string UID = 3;</code>
+       */
+      public Builder addUID(
+          java.lang.String value) {
         if (value == null) {
-          throw new NullPointerException();
-        }
-        if (result.uID_.isEmpty()) {
-          result.uID_ = new java.util.ArrayList<java.lang.String>();
-        }
-        result.uID_.add(value);
+    throw new NullPointerException();
+  }
+  ensureUIDIsMutable();
+        uID_.add(value);
+        onChanged();
         return this;
       }
-      
-      public Builder addAllUID(java.lang.Iterable<? extends java.lang.String> values) {
-        if (result.uID_.isEmpty()) {
-          result.uID_ = new java.util.ArrayList<java.lang.String>();
-        }
-        super.addAll(values, result.uID_);
+      /**
+       * <code>repeated string UID = 3;</code>
+       */
+      public Builder addAllUID(
+          java.lang.Iterable<java.lang.String> values) {
+        ensureUIDIsMutable();
+        super.addAll(values, uID_);
+        onChanged();
         return this;
       }
-      
+      /**
+       * <code>repeated string UID = 3;</code>
+       */
       public Builder clearUID() {
-        result.uID_ = java.util.Collections.emptyList();
+        uID_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+        bitField0_ = (bitField0_ & ~0x00000004);
+        onChanged();
+        return this;
+      }
+      /**
+       * <code>repeated string UID = 3;</code>
+       */
+      public Builder addUIDBytes(
+          com.google.protobuf.ByteString value) {
+        if (value == null) {
+    throw new NullPointerException();
+  }
+  ensureUIDIsMutable();
+        uID_.add(value);
+        onChanged();
         return this;
       }
-      
-      // @@protoc_insertion_point(builder_scope:protobuf.List)
+
+      // @@protoc_insertion_point(builder_scope:org.apache.accumulo.examples.wikisearch.protobuf.List)
     }
-    
+
     static {
       defaultInstance = new List(true);
-      org.apache.accumulo.examples.wikisearch.protobuf.Uid.internalForceInit();
       defaultInstance.initFields();
     }
-    
-    // @@protoc_insertion_point(class_scope:protobuf.List)
+
+    // @@protoc_insertion_point(class_scope:org.apache.accumulo.examples.wikisearch.protobuf.List)
   }
-  
-  private static com.google.protobuf.Descriptors.Descriptor internal_static_protobuf_List_descriptor;
-  private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_protobuf_List_fieldAccessorTable;
-  
-  public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() {
+
+  private static com.google.protobuf.Descriptors.Descriptor
+    internal_static_org_apache_accumulo_examples_wikisearch_protobuf_List_descriptor;
+  private static
+    com.google.protobuf.GeneratedMessage.FieldAccessorTable
+      internal_static_org_apache_accumulo_examples_wikisearch_protobuf_List_fieldAccessorTable;
+
+  public static com.google.protobuf.Descriptors.FileDescriptor
+      getDescriptor() {
     return descriptor;
   }
-  
-  private static com.google.protobuf.Descriptors.FileDescriptor descriptor;
+  private static com.google.protobuf.Descriptors.FileDescriptor
+      descriptor;
   static {
-    java.lang.String[] descriptorData = {"\n\tUid.proto\022\010protobuf\"2\n\004List\022\016\n\006IGNORE\030"
-        + "\001 \002(\010\022\r\n\005COUNT\030\002 \002(\004\022\013\n\003UID\030\003 \003(\tB\014\n\010pro" + "tobufH\001"};
-    com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
-      public com.google.protobuf.ExtensionRegistry assignDescriptors(com.google.protobuf.Descriptors.FileDescriptor root) {
-        descriptor = root;
-        internal_static_protobuf_List_descriptor = getDescriptor().getMessageTypes().get(0);
-        internal_static_protobuf_List_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable(
-            internal_static_protobuf_List_descriptor, new java.lang.String[] {"IGNORE", "COUNT", "UID",}, org.apache.accumulo.examples.wikisearch.protobuf.Uid.List.class,
-            org.apache.accumulo.examples.wikisearch.protobuf.Uid.List.Builder.class);
-        return null;
-      }
+    java.lang.String[] descriptorData = {
+      "\n\tUid.proto\0220org.apache.accumulo.example" +
+      "s.wikisearch.protobuf\"2\n\004List\022\016\n\006IGNORE\030" +
+      "\001 \002(\010\022\r\n\005COUNT\030\002 \002(\004\022\013\n\003UID\030\003 \003(\tB4\n0org" +
+      ".apache.accumulo.examples.wikisearch.pro" +
+      "tobufH\001"
     };
-    com.google.protobuf.Descriptors.FileDescriptor.internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] {},
-        assigner);
+    com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
+      new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
+        public com.google.protobuf.ExtensionRegistry assignDescriptors(
+            com.google.protobuf.Descriptors.FileDescriptor root) {
+          descriptor = root;
+          internal_static_org_apache_accumulo_examples_wikisearch_protobuf_List_descriptor =
+            getDescriptor().getMessageTypes().get(0);
+          internal_static_org_apache_accumulo_examples_wikisearch_protobuf_List_fieldAccessorTable = new
+            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
+              internal_static_org_apache_accumulo_examples_wikisearch_protobuf_List_descriptor,
+              new java.lang.String[] { "IGNORE", "COUNT", "UID", });
+          return null;
+        }
+      };
+    com.google.protobuf.Descriptors.FileDescriptor
+      .internalBuildGeneratedFileFrom(descriptorData,
+        new com.google.protobuf.Descriptors.FileDescriptor[] {
+        }, assigner);
   }
-  
-  public static void internalForceInit() {}
-  
+
   // @@protoc_insertion_point(outer_class_scope)
 }

http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/e32632ca/ingest/src/main/protobuf/TermWeight.proto
----------------------------------------------------------------------
diff --git a/ingest/src/main/protobuf/TermWeight.proto b/ingest/src/main/protobuf/TermWeight.proto
index 41ae188..3d67c31 100644
--- a/ingest/src/main/protobuf/TermWeight.proto
+++ b/ingest/src/main/protobuf/TermWeight.proto
@@ -17,9 +17,9 @@
 // compile extra builder util with java accumulo.data.protobuf.builder.ProtoBufBuilder -d ../java accumulo.data.protobuf.UidList
 //      classpath for compile command should include ../../../target/classes and protobuf-java-2.2.0.jar
 
-package protobuf;
+package org.apache.accumulo.examples.wikisearch.protobuf;
 
-option java_package = "protobuf";
+option java_package = "org.apache.accumulo.examples.wikisearch.protobuf";
 option optimize_for = SPEED;
 
 message Info {

http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/e32632ca/ingest/src/main/protobuf/Uid.proto
----------------------------------------------------------------------
diff --git a/ingest/src/main/protobuf/Uid.proto b/ingest/src/main/protobuf/Uid.proto
index 30aa446..db78b5d 100644
--- a/ingest/src/main/protobuf/Uid.proto
+++ b/ingest/src/main/protobuf/Uid.proto
@@ -17,9 +17,9 @@
 // compile extra builder util with java accumulo.data.protobuf.builder.ProtoBufBuilder -d ../java accumulo.data.protobuf.UidList
 //      classpath for compile command should include ../../../target/classes and protobuf-java-2.2.0.jar
 
-package protobuf;
+package org.apache.accumulo.examples.wikisearch.protobuf;
 
-option java_package = "protobuf";
+option java_package = "org.apache.accumulo.examples.wikisearch.protobuf";
 option optimize_for = SPEED;
 
 message List {

http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/e32632ca/pom.xml
----------------------------------------------------------------------
diff --git a/pom.xml b/pom.xml
index 2e3cb3f..bebc1cf 100644
--- a/pom.xml
+++ b/pom.xml
@@ -21,12 +21,12 @@
   <parent>
       <groupId>org.apache</groupId>
       <artifactId>apache</artifactId>
-      <version>14</version>
+      <version>18</version>
   </parent>
 
   <groupId>org.apache.accumulo</groupId>
   <artifactId>accumulo-wikisearch</artifactId>
-  <version>1.5.0</version>
+  <version>1.8.0</version>
 
   <packaging>pom</packaging>
   <name>accumulo-wikisearch</name>
@@ -44,7 +44,7 @@
     <version.commons-lang>2.4</version.commons-lang>
     <version.ejb-spec-api>1.0.1.Final</version.ejb-spec-api>
     <version.guava>11.0.2</version.guava>
-    <version.hadoop>1.0.4</version.hadoop>
+    <version.hadoop>2.6.4</version.hadoop>
     <version.jaxrs>2.1.0.GA</version.jaxrs>
     <version.junit>4.11</version.junit>
     <version.kryo>1.04</version.kryo>
@@ -54,8 +54,8 @@
     <version.lucene-analyzers>3.0.2</version.lucene-analyzers>
     <version.lucene-wikipedia>3.0.2</version.lucene-wikipedia>
     <version.minlog>1.2</version.minlog>
-    <version.protobuf>2.3.0</version.protobuf>
-    <version.thrift>0.9.0</version.thrift>
+    <version.protobuf>2.5.0</version.protobuf>
+    <version.thrift>0.9.3</version.thrift>
     <version.zookeeper>3.3.1</version.zookeeper>
   </properties>
   <dependencyManagement>
@@ -76,6 +76,16 @@
       <version>${version.collections}</version>
     </dependency>
     <dependency>
+      <groupId>commons-logging</groupId>
+      <artifactId>commons-logging</artifactId>
+      <version>1.1.1</version>
+    </dependency>
+    <dependency>
+      <groupId>org.apache.commons</groupId>
+      <artifactId>commons-math3</artifactId>
+      <version>3.6.1</version>
+    </dependency>
+    <dependency>
       <groupId>junit</groupId>
       <artifactId>junit</artifactId>
       <version>${version.junit}</version>
@@ -206,6 +216,21 @@
             </exclusion>
         </exclusions>
       </dependency>
+      <dependency>
+        <groupId>org.slf4j</groupId>
+        <artifactId>slf4j-api</artifactId>
+        <version>1.7.21</version>
+      </dependency>
+      <dependency>
+        <groupId>org.codehaus.jackson</groupId>
+        <artifactId>jackson-core-asl</artifactId>
+        <version>1.9.13</version>
+      </dependency>
+      <dependency>
+        <groupId>org.codehaus.jackson</groupId>
+        <artifactId>jackson-mapper-asl</artifactId>
+        <version>1.9.13</version>
+      </dependency>
     </dependencies>
   </dependencyManagement>
   <repositories>
@@ -351,73 +376,4 @@
       </plugin>
     </plugins>
   </build>
-  <profiles>
-    <profile>
-      <!-- profile for building against Hadoop 1.0.x
-      Activate by not specifying hadoop.profile -->
-      <id>hadoop-1.0</id>
-      <activation>
-        <property>
-          <name>!hadoop.profile</name>
-        </property>
-      </activation>
-      <build>
-        <plugins>
-          <plugin>
-            <groupId>org.codehaus.mojo</groupId>
-            <artifactId>build-helper-maven-plugin</artifactId>
-            <version>1.8</version>
-            <executions>
-              <execution>
-                <id>add-test-source</id>
-                <goals>
-                  <goal>add-test-source</goal>
-                </goals>
-                <phase>generate-test-sources</phase>
-                <configuration>
-                  <sources>
-                    <source>src/test/hadoop1</source>
-                  </sources>
-                </configuration>
-              </execution>
-            </executions>
-          </plugin>
-        </plugins>
-      </build>
-    </profile>
-    <profile>
-      <!-- profile for building against Hadoop 2.0.x
-      Activate using: mvn -Dhadoop.profile=2.0 -->
-      <id>hadoop-2.0</id>
-      <activation>
-        <property>
-          <name>hadoop.profile</name>
-          <value>2.0</value>
-        </property>
-      </activation>
-      <build>
-        <plugins>
-          <plugin>
-            <groupId>org.codehaus.mojo</groupId>
-            <artifactId>build-helper-maven-plugin</artifactId>
-            <version>1.8</version>
-            <executions>
-              <execution>
-                <id>add-test-source</id>
-                <goals>
-                  <goal>add-test-source</goal>
-                </goals>
-                <phase>generate-test-sources</phase>
-                <configuration>
-                  <sources>
-                    <source>src/test/hadoop2</source>
-                  </sources>
-                </configuration>
-              </execution>
-            </executions>
-          </plugin>
-        </plugins>
-      </build>
-    </profile>
-  </profiles>
 </project>

http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/e32632ca/query-war/pom.xml
----------------------------------------------------------------------
diff --git a/query-war/pom.xml b/query-war/pom.xml
index 3ee0bc2..45a1f1b 100644
--- a/query-war/pom.xml
+++ b/query-war/pom.xml
@@ -20,7 +20,7 @@
   <parent>
     <groupId>org.apache.accumulo</groupId>
     <artifactId>accumulo-wikisearch</artifactId>
-    <version>1.5.0</version>
+    <version>1.8.0</version>
   </parent>
 
   <artifactId>wikisearch-query-war</artifactId>

http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/e32632ca/query/pom.xml
----------------------------------------------------------------------
diff --git a/query/pom.xml b/query/pom.xml
index e779945..eeb2337 100644
--- a/query/pom.xml
+++ b/query/pom.xml
@@ -20,7 +20,7 @@
   <parent>
     <groupId>org.apache.accumulo</groupId>
     <artifactId>accumulo-wikisearch</artifactId>
-    <version>1.5.0</version>
+    <version>1.8.0</version>
   </parent>
   <artifactId>wikisearch-query</artifactId>
   <packaging>ejb</packaging>
@@ -51,6 +51,11 @@
       <artifactId>commons-collections</artifactId>
     </dependency>
     <dependency>
+      <groupId>commons-configuration</groupId>
+      <artifactId>commons-configuration</artifactId>
+      <version>1.6</version>
+    </dependency>
+    <dependency>
       <groupId>commons-lang</groupId>
       <artifactId>commons-lang</artifactId>
     </dependency>
@@ -83,21 +88,9 @@
       <scope>runtime</scope>
     </dependency>
     <dependency>
-      <groupId>commons-configuration</groupId>
-      <artifactId>commons-configuration</artifactId>
-      <version>1.6</version>
-      <exclusions>
-          <exclusion>
-              <groupId>commons-logging</groupId>
-              <artifactId>commons-logging</artifactId>
-          </exclusion>
-      </exclusions>
-      <scope>runtime</scope>
-    </dependency>
-    <dependency>
       <groupId>commons-io</groupId>
       <artifactId>commons-io</artifactId>
-      <version>2.1</version>
+      <version>2.4</version>
       <scope>runtime</scope>
     </dependency>
     <dependency>
@@ -142,7 +135,7 @@
               <outputDirectory>lib</outputDirectory>
               <!-- just grab the non-provided runtime dependencies -->
               <!-- XXX we include guava at the same version as hadoop 2 provides so that we have it on hadoop 1 -->
-              <includeArtifactIds>commons-io,commons-configuration,commons-lang,commons-codec,protobuf-java,libthrift,zookeeper,hadoop-core,commons-jexl,guava,kryo,asm,minlog,reflectasm,wikisearch-ingest,accumulo-core,accumulo-fate,accumulo-trace</includeArtifactIds>
+              <includeArtifactIds>commons-io,commons-configuration,commons-lang,commons-codec,protobuf-java,libthrift,zookeeper,hadoop-client,commons-jexl,guava,kryo,asm,minlog,reflectasm,wikisearch-ingest,accumulo-core,accumulo-fate,accumulo-trace</includeArtifactIds>
               <excludeTransitive>true</excludeTransitive>
             </configuration>
           </execution>
@@ -178,76 +171,4 @@
       </plugin>
     </plugins>
   </build>
-  <profiles>
-    <!-- profile for building against Hadoop 1.0.x
-    Activate by not specifying hadoop.profile -->
-    <profile>
-      <id>hadoop-1.0</id>
-      <activation>
-        <property>
-          <name>!hadoop.profile</name>
-        </property>
-      </activation>
-      <dependencies>
-        <dependency>
-          <groupId>org.apache.hadoop</groupId>
-          <artifactId>hadoop-core</artifactId>
-          <version>${version.hadoop}</version>
-          <exclusions>
-            <exclusion>
-                <groupId>commons-logging</groupId>
-                <artifactId>commons-logging</artifactId>
-            </exclusion>
-          </exclusions>
-        </dependency>
-      </dependencies>
-    </profile>
-    <!-- profile for building against Hadoop 2.0.x
-    Activate using: mvn -Dhadoop.profile=2.0 -->
-    <profile>
-      <id>hadoop-2.0</id>
-      <activation>
-        <property>
-          <name>hadoop.profile</name>
-          <value>2.0</value>
-        </property>
-      </activation>
-      <dependencies>
-        <dependency>
-          <groupId>org.apache.hadoop</groupId>
-          <artifactId>hadoop-client</artifactId>
-          <version>${version.hadoop}</version>
-          <exclusions>
-            <exclusion>
-              <groupId>org.codehaus.jackson</groupId>
-              <artifactId>jackson-mapper-asl</artifactId>
-            </exclusion>
-          </exclusions>
-        </dependency>
-      </dependencies>
-      <build>
-        <plugins>
-          <plugin>
-            <groupId>org.apache.maven.plugins</groupId>
-            <artifactId>maven-dependency-plugin</artifactId>
-            <executions>
-              <execution>
-                <id>copy-dependencies</id>
-                <goals>
-                  <goal>copy-dependencies</goal>
-                </goals>
-                <phase>prepare-package</phase>
-                <configuration>
-                  <outputDirectory>lib</outputDirectory>
-                  <!-- just grab the non-provided runtime dependencies -->
-                  <includeArtifactIds>commons-io,commons-configuration,commons-lang,commons-codec,protobuf-java,libthrift,zookeeper,hadoop-client,hadoop-common,hadoop-hdfs,commons-jexl,guava,kryo,asm,minlog,reflectasm,wikisearch-ingest,accumulo-core,accumulo-fate,accumulo-trace</includeArtifactIds>
-                  <excludeTransitive>false</excludeTransitive>
-                </configuration>
-              </execution>
-            </executions>
-          </plugin>
-        </plugins>
-      </build>
-    </profile>
-  </profiles>
 </project>

http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/e32632ca/query/src/main/java/org/apache/accumulo/examples/wikisearch/iterator/DefaultIteratorEnvironment.java
----------------------------------------------------------------------
diff --git a/query/src/main/java/org/apache/accumulo/examples/wikisearch/iterator/DefaultIteratorEnvironment.java b/query/src/main/java/org/apache/accumulo/examples/wikisearch/iterator/DefaultIteratorEnvironment.java
index 6b58d08..f05304a 100644
--- a/query/src/main/java/org/apache/accumulo/examples/wikisearch/iterator/DefaultIteratorEnvironment.java
+++ b/query/src/main/java/org/apache/accumulo/examples/wikisearch/iterator/DefaultIteratorEnvironment.java
@@ -18,6 +18,7 @@ package org.apache.accumulo.examples.wikisearch.iterator;
 
 import java.io.IOException;
 
+import org.apache.accumulo.core.client.sample.SamplerConfiguration;
 import org.apache.accumulo.core.conf.AccumuloConfiguration;
 import org.apache.accumulo.core.data.Key;
 import org.apache.accumulo.core.data.Value;
@@ -25,6 +26,7 @@ import org.apache.accumulo.core.iterators.IteratorEnvironment;
 import org.apache.accumulo.core.iterators.IteratorUtil.IteratorScope;
 import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
 import org.apache.accumulo.core.iterators.system.MapFileIterator;
+import org.apache.accumulo.core.security.Authorizations;
 import org.apache.accumulo.core.util.CachedConfiguration;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FileSystem;
@@ -70,5 +72,25 @@ public class DefaultIteratorEnvironment implements IteratorEnvironment {
   public void registerSideChannel(SortedKeyValueIterator<Key,Value> iter) {
     throw new UnsupportedOperationException();
   }
+
+@Override
+public Authorizations getAuthorizations() {
+	throw new UnsupportedOperationException();
+}
+
+@Override
+public IteratorEnvironment cloneWithSamplingEnabled() {
+	throw new UnsupportedOperationException();
+}
+
+@Override
+public boolean isSamplingEnabled() {
+	throw new UnsupportedOperationException();
+}
+
+@Override
+public SamplerConfiguration getSamplerConfiguration() {
+	throw new UnsupportedOperationException();
+}
   
 }

http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/e32632ca/query/src/main/java/org/apache/accumulo/examples/wikisearch/query/Query.java
----------------------------------------------------------------------
diff --git a/query/src/main/java/org/apache/accumulo/examples/wikisearch/query/Query.java b/query/src/main/java/org/apache/accumulo/examples/wikisearch/query/Query.java
index 024a865..62aa4e2 100644
--- a/query/src/main/java/org/apache/accumulo/examples/wikisearch/query/Query.java
+++ b/query/src/main/java/org/apache/accumulo/examples/wikisearch/query/Query.java
@@ -43,6 +43,7 @@ import javax.xml.transform.stream.StreamSource;
 import org.apache.accumulo.core.client.Connector;
 import org.apache.accumulo.core.client.Instance;
 import org.apache.accumulo.core.client.ZooKeeperInstance;
+import org.apache.accumulo.core.client.security.tokens.PasswordToken;
 import org.apache.accumulo.examples.wikisearch.logic.ContentLogic;
 import org.apache.accumulo.examples.wikisearch.logic.QueryLogic;
 import org.apache.accumulo.examples.wikisearch.sample.Results;
@@ -179,7 +180,7 @@ public class Query implements IQuery {
     Instance instance = new ZooKeeperInstance(this.instanceName, this.zooKeepers);
     try {
       log.info("Connecting to [instanceName = " + this.instanceName + ", zookeepers = " + this.zooKeepers + ", username = " + this.username + "].");
-      connector = instance.getConnector(this.username, this.password.getBytes());
+      connector = instance.getConnector(this.username, new PasswordToken(this.password.getBytes()));
     } catch (Exception e) {
       throw new EJBException("Error getting connector from instance", e);
     }
@@ -212,7 +213,7 @@ public class Query implements IQuery {
     Instance instance = new ZooKeeperInstance(this.instanceName, this.zooKeepers);
     try {
       log.info("Connecting to [instanceName = " + this.instanceName + ", zookeepers = " + this.zooKeepers + ", username = " + this.username + "].");
-      connector = instance.getConnector(this.username, this.password.getBytes());
+      connector = instance.getConnector(this.username, new PasswordToken(this.password.getBytes()));
     } catch (Exception e) {
       throw new EJBException("Error getting connector from instance", e);
     }


[2/2] accumulo-wikisearch git commit: ACCUMULO-2446: many updates for Accumulo 1.8.0

Posted by mm...@apache.org.
ACCUMULO-2446: many updates for Accumulo 1.8.0


Project: http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/commit/e32632ca
Tree: http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/tree/e32632ca
Diff: http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/diff/e32632ca

Branch: refs/heads/1.8
Commit: e32632cafc9a15cce716c7cb43423a4e48e0d2c3
Parents: 9c30660
Author: Mike Miller <mm...@apache.org>
Authored: Thu Dec 1 15:18:30 2016 -0500
Committer: Mike Miller <mm...@apache.org>
Committed: Tue Dec 13 13:22:02 2016 -0500

----------------------------------------------------------------------
 .gitignore                                      |   4 +
 ingest/bin/ingest.sh                            |   2 +-
 ingest/conf/wikipedia.xml.uno                   |  43 +
 ingest/pom.xml                                  |  57 +-
 .../wikisearch/ingest/WikipediaIngester.java    |  13 +-
 .../ingest/WikipediaPartitionedIngester.java    |  12 +-
 .../output/BufferingRFileRecordWriter.java      |  47 +-
 .../output/SortingRFileOutputFormat.java        |   9 +-
 .../wikisearch/protobuf/TermWeight.java         | 721 ++++++++++------
 .../examples/wikisearch/protobuf/Uid.java       | 838 ++++++++++++-------
 ingest/src/main/protobuf/TermWeight.proto       |   4 +-
 ingest/src/main/protobuf/Uid.proto              |   4 +-
 pom.xml                                         | 104 +--
 query-war/pom.xml                               |   2 +-
 query/pom.xml                                   |  95 +--
 .../iterator/DefaultIteratorEnvironment.java    |  22 +
 .../examples/wikisearch/query/Query.java        |   5 +-
 17 files changed, 1168 insertions(+), 814 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/e32632ca/.gitignore
----------------------------------------------------------------------
diff --git a/.gitignore b/.gitignore
index 698a832..8a09ac0 100644
--- a/.gitignore
+++ b/.gitignore
@@ -2,3 +2,7 @@
 .idea
 **/*.iml
 **/lib
+.project
+.settings/
+.classpath
+

http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/e32632ca/ingest/bin/ingest.sh
----------------------------------------------------------------------
diff --git a/ingest/bin/ingest.sh b/ingest/bin/ingest.sh
index 73d582d..f9f3495 100755
--- a/ingest/bin/ingest.sh
+++ b/ingest/bin/ingest.sh
@@ -38,7 +38,7 @@ LIBJARS=`echo $CLASSPATH | sed 's/^://' | sed 's/:/,/g'`
 #
 # Map/Reduce job
 #
-JAR=$SCRIPT_DIR/../lib/wikisearch-ingest-1.5.0.jar
+JAR=$SCRIPT_DIR/../lib/wikisearch-ingest-*.jar
 CONF=$SCRIPT_DIR/../conf/wikipedia.xml
 HDFS_DATA_DIR=$1
 export HADOOP_CLASSPATH=$CLASSPATH

http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/e32632ca/ingest/conf/wikipedia.xml.uno
----------------------------------------------------------------------
diff --git a/ingest/conf/wikipedia.xml.uno b/ingest/conf/wikipedia.xml.uno
new file mode 100644
index 0000000..093551c
--- /dev/null
+++ b/ingest/conf/wikipedia.xml.uno
@@ -0,0 +1,43 @@
+<?xml version="1.0"?>
+<!--
+  Licensed to the Apache Software Foundation (ASF) under one or more
+  contributor license agreements.  See the NOTICE file distributed with
+  this work for additional information regarding copyright ownership.
+  The ASF licenses this file to You under the Apache License, Version 2.0
+  (the "License"); you may not use this file except in compliance with
+  the License.  You may obtain a copy of the License at
+
+      http://www.apache.org/licenses/LICENSE-2.0
+
+  Unless required by applicable law or agreed to in writing, software
+  distributed under the License is distributed on an "AS IS" BASIS,
+  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+  See the License for the specific language governing permissions and
+  limitations under the License.
+-->
+<configuration>
+  <property>
+    <name>wikipedia.accumulo.zookeepers</name>
+    <value>localhost:2181</value>
+  </property>
+  <property>
+    <name>wikipedia.accumulo.instance_name</name>
+    <value>uno</value>
+  </property>
+  <property>
+    <name>wikipedia.accumulo.user</name>
+    <value>root</value>
+  </property>
+  <property>
+    <name>wikipedia.accumulo.password</name>
+    <value>secret</value>
+  </property>
+  <property>
+    <name>wikipedia.accumulo.table</name>
+    <value>wikipedia</value>
+  </property>
+  <property>
+    <name>wikipedia.ingest.partitions</name>
+    <value>1</value>
+  </property>
+</configuration>

http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/e32632ca/ingest/pom.xml
----------------------------------------------------------------------
diff --git a/ingest/pom.xml b/ingest/pom.xml
index 1d17e27..426cff2 100644
--- a/ingest/pom.xml
+++ b/ingest/pom.xml
@@ -20,7 +20,7 @@
   <parent>
     <groupId>org.apache.accumulo</groupId>
     <artifactId>accumulo-wikisearch</artifactId>
-    <version>1.5.0</version>
+    <version>1.8.0</version>
   </parent>
   <artifactId>wikisearch-ingest</artifactId>
   <name>wikisearch-ingest</name>
@@ -48,6 +48,12 @@
     <dependency>
       <groupId>org.apache.accumulo</groupId>
       <artifactId>accumulo-core</artifactId>
+        <exclusions>
+          <exclusion>
+            <groupId>commons-digester</groupId>
+            <artifactId>commons-digester</artifactId>
+          </exclusion>
+        </exclusions>
     </dependency>
     <dependency>
       <groupId>org.apache.lucene</groupId>
@@ -101,53 +107,4 @@
       </plugin>
     </plugins>
   </build>
-  <profiles>
-    <!-- profile for building against Hadoop 1.0.x
-    Activate by not specifying hadoop.profile -->
-    <profile>
-      <id>hadoop-1.0</id>
-      <activation>
-        <property>
-          <name>!hadoop.profile</name>
-        </property>
-      </activation>
-      <dependencies>
-        <dependency>
-          <groupId>org.apache.hadoop</groupId>
-          <artifactId>hadoop-core</artifactId>
-          <version>${version.hadoop}</version>
-          <exclusions>
-              <exclusion>
-                  <groupId>commons-logging</groupId>
-                  <artifactId>commons-logging</artifactId>
-              </exclusion>
-          </exclusions>
-        </dependency>
-      </dependencies>
-    </profile>
-    <!-- profile for building against Hadoop 2.0.x
-    Activate using: mvn -Dhadoop.profile=2.0 -->
-    <profile>
-      <id>hadoop-2.0</id>
-      <activation>
-        <property>
-          <name>hadoop.profile</name>
-          <value>2.0</value>
-        </property>
-      </activation>
-      <dependencies>
-        <dependency>
-          <groupId>org.apache.hadoop</groupId>
-          <artifactId>hadoop-client</artifactId>
-          <version>${version.hadoop}</version>
-          <exclusions>
-            <exclusion>
-              <groupId>org.codehaus.jackson</groupId>
-              <artifactId>jackson-mapper-asl</artifactId>
-            </exclusion>
-          </exclusions>
-        </dependency>
-      </dependencies>
-    </profile>
-  </profiles>
 </project>

http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/e32632ca/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/ingest/WikipediaIngester.java
----------------------------------------------------------------------
diff --git a/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/ingest/WikipediaIngester.java b/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/ingest/WikipediaIngester.java
index d4fa1c6..1a495ed 100644
--- a/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/ingest/WikipediaIngester.java
+++ b/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/ingest/WikipediaIngester.java
@@ -29,13 +29,16 @@ import java.util.regex.Pattern;
 
 import org.apache.accumulo.core.client.AccumuloException;
 import org.apache.accumulo.core.client.AccumuloSecurityException;
+import org.apache.accumulo.core.client.ClientConfiguration;
 import org.apache.accumulo.core.client.Connector;
 import org.apache.accumulo.core.client.IteratorSetting;
 import org.apache.accumulo.core.client.IteratorSetting.Column;
 import org.apache.accumulo.core.client.TableExistsException;
 import org.apache.accumulo.core.client.TableNotFoundException;
+import org.apache.accumulo.core.client.ClientConfiguration.ClientProperty;
 import org.apache.accumulo.core.client.admin.TableOperations;
 import org.apache.accumulo.core.client.mapreduce.AccumuloOutputFormat;
+import org.apache.accumulo.core.client.security.tokens.PasswordToken;
 import org.apache.accumulo.core.data.Mutation;
 import org.apache.accumulo.core.iterators.IteratorUtil.IteratorScope;
 import org.apache.accumulo.core.iterators.user.SummingCombiner;
@@ -135,9 +138,9 @@ public class WikipediaIngester extends Configured implements Tool {
     conf.set("mapred.map.tasks.speculative.execution", "false");
     
     String tablename = WikipediaConfiguration.getTableName(conf);
-    
-    String zookeepers = WikipediaConfiguration.getZookeepers(conf);
-    String instanceName = WikipediaConfiguration.getInstanceName(conf);
+    ClientConfiguration clientConfig = new ClientConfiguration();
+    clientConfig.setProperty(ClientProperty.INSTANCE_NAME, WikipediaConfiguration.getInstanceName(conf));
+    clientConfig.setProperty(ClientProperty.INSTANCE_ZK_HOST, WikipediaConfiguration.getZookeepers(conf));
     
     String user = WikipediaConfiguration.getUser(conf);
     byte[] password = WikipediaConfiguration.getPassword(conf);
@@ -168,8 +171,8 @@ public class WikipediaIngester extends Configured implements Tool {
     job.setMapOutputKeyClass(Text.class);
     job.setMapOutputValueClass(Mutation.class);
     job.setOutputFormatClass(AccumuloOutputFormat.class);
-    AccumuloOutputFormat.setOutputInfo(job.getConfiguration(), user, password, true, tablename);
-    AccumuloOutputFormat.setZooKeeperInstance(job.getConfiguration(), instanceName, zookeepers);
+    AccumuloOutputFormat.setConnectorInfo(job, user, new PasswordToken(password));
+    AccumuloOutputFormat.setZooKeeperInstance(job, clientConfig);
     
     return job.waitForCompletion(true) ? 0 : 1;
   }

http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/e32632ca/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/ingest/WikipediaPartitionedIngester.java
----------------------------------------------------------------------
diff --git a/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/ingest/WikipediaPartitionedIngester.java b/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/ingest/WikipediaPartitionedIngester.java
index 59035dc..841f169 100644
--- a/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/ingest/WikipediaPartitionedIngester.java
+++ b/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/ingest/WikipediaPartitionedIngester.java
@@ -29,13 +29,16 @@ import java.util.regex.Pattern;
 
 import org.apache.accumulo.core.client.AccumuloException;
 import org.apache.accumulo.core.client.AccumuloSecurityException;
+import org.apache.accumulo.core.client.ClientConfiguration;
 import org.apache.accumulo.core.client.Connector;
 import org.apache.accumulo.core.client.IteratorSetting;
 import org.apache.accumulo.core.client.IteratorSetting.Column;
 import org.apache.accumulo.core.client.TableExistsException;
 import org.apache.accumulo.core.client.TableNotFoundException;
+import org.apache.accumulo.core.client.ClientConfiguration.ClientProperty;
 import org.apache.accumulo.core.client.admin.TableOperations;
 import org.apache.accumulo.core.client.mapreduce.AccumuloOutputFormat;
+import org.apache.accumulo.core.client.security.tokens.PasswordToken;
 import org.apache.accumulo.core.data.Mutation;
 import org.apache.accumulo.core.iterators.IteratorUtil.IteratorScope;
 import org.apache.accumulo.core.iterators.user.SummingCombiner;
@@ -236,12 +239,13 @@ public class WikipediaPartitionedIngester extends Configured implements Tool {
       SortingRFileOutputFormat.setPathName(ingestConf, WikipediaConfiguration.bulkIngestDir(ingestConf));
     } else {
       ingestJob.setOutputFormatClass(AccumuloOutputFormat.class);
-      String zookeepers = WikipediaConfiguration.getZookeepers(ingestConf);
-      String instanceName = WikipediaConfiguration.getInstanceName(ingestConf);
+      ClientConfiguration clientConfig = new ClientConfiguration();
+      clientConfig.setProperty(ClientProperty.INSTANCE_NAME, WikipediaConfiguration.getInstanceName(ingestConf));
+      clientConfig.setProperty(ClientProperty.INSTANCE_ZK_HOST, WikipediaConfiguration.getZookeepers(ingestConf));
       String user = WikipediaConfiguration.getUser(ingestConf);
       byte[] password = WikipediaConfiguration.getPassword(ingestConf);
-      AccumuloOutputFormat.setOutputInfo(ingestJob.getConfiguration(), user, password, true, tablename);
-      AccumuloOutputFormat.setZooKeeperInstance(ingestJob.getConfiguration(), instanceName, zookeepers);
+      AccumuloOutputFormat.setConnectorInfo(ingestJob, user, new PasswordToken(password));
+      AccumuloOutputFormat.setZooKeeperInstance(ingestJob, clientConfig);
     }
     
     return ingestJob.waitForCompletion(true) ? 0 : 1;

http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/e32632ca/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/output/BufferingRFileRecordWriter.java
----------------------------------------------------------------------
diff --git a/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/output/BufferingRFileRecordWriter.java b/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/output/BufferingRFileRecordWriter.java
index 9b663de..aa6e357 100644
--- a/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/output/BufferingRFileRecordWriter.java
+++ b/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/output/BufferingRFileRecordWriter.java
@@ -22,13 +22,18 @@ import java.util.Map;
 import java.util.Map.Entry;
 import java.util.TreeMap;
 
+import org.apache.accumulo.core.client.AccumuloException;
+import org.apache.accumulo.core.client.AccumuloSecurityException;
+import org.apache.accumulo.core.client.BatchWriter;
+import org.apache.accumulo.core.client.BatchWriterConfig;
+import org.apache.accumulo.core.client.Connector;
+import org.apache.accumulo.core.client.TableNotFoundException;
 import org.apache.accumulo.core.conf.AccumuloConfiguration;
 import org.apache.accumulo.core.data.ColumnUpdate;
 import org.apache.accumulo.core.data.Key;
 import org.apache.accumulo.core.data.Mutation;
 import org.apache.accumulo.core.data.Value;
-import org.apache.accumulo.core.file.FileSKVWriter;
-import org.apache.accumulo.core.file.rfile.RFileOperations;
+import org.apache.accumulo.examples.wikisearch.ingest.WikipediaConfiguration;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.io.Text;
@@ -37,12 +42,7 @@ import org.apache.hadoop.mapreduce.TaskAttemptContext;
 
 final class BufferingRFileRecordWriter extends RecordWriter<Text,Mutation> {
   private final long maxSize;
-  private final AccumuloConfiguration acuconf;
   private final Configuration conf;
-  private final String filenamePrefix;
-  private final String taskID;
-  private final FileSystem fs;
-  private int fileCount = 0;
   private long size;
   
   private Map<Text,TreeMap<Key,Value>> buffers = new HashMap<Text,TreeMap<Key,Value>>();
@@ -79,31 +79,32 @@ final class BufferingRFileRecordWriter extends RecordWriter<Text,Mutation> {
     if (buffer.size() == 0)
       return;
     
-    String file = filenamePrefix + "/" + tablename + "/" + taskID + "_" + (fileCount++) + ".rf";
+    Connector conn;
+	try {		
+	  conn = WikipediaConfiguration.getConnector(conf);
+      BatchWriterConfig bwconfig = new BatchWriterConfig();
+      BatchWriter writer = conn.createBatchWriter(tablename.toString(), bwconfig);
+      for (Entry<Key,Value> e : buffer.entrySet()) {
+        Key k = e.getKey();
+    	Mutation m = new Mutation();
+    	m.put(k.getColumnFamily(), k.getColumnQualifier(), e.getValue());
+        writer.addMutation(m);
+      }
+      writer.close();
+	} catch (AccumuloException | AccumuloSecurityException | TableNotFoundException e1) {
+	  // TODO Auto-generated catch block
+	  e1.printStackTrace();
+	}    
     // TODO get the table configuration for the given table?
-    FileSKVWriter writer = RFileOperations.getInstance().openWriter(file, fs, conf, acuconf);
-    
-    // forget locality groups for now, just write everything to the default
-    writer.startDefaultLocalityGroup();
-    
-    for (Entry<Key,Value> e : buffer.entrySet()) {
-      writer.append(e.getKey(), e.getValue());
-    }
-    
-    writer.close();
     
     size -= bufferSize;
     buffer.clear();
     bufferSizes.put(tablename, 0l);
   }
   
-  BufferingRFileRecordWriter(long maxSize, AccumuloConfiguration acuconf, Configuration conf, String filenamePrefix, String taskID, FileSystem fs) {
+  BufferingRFileRecordWriter(long maxSize, Configuration conf) {
     this.maxSize = maxSize;
-    this.acuconf = acuconf;
     this.conf = conf;
-    this.filenamePrefix = filenamePrefix;
-    this.taskID = taskID;
-    this.fs = fs;
   }
   
   @Override

http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/e32632ca/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/output/SortingRFileOutputFormat.java
----------------------------------------------------------------------
diff --git a/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/output/SortingRFileOutputFormat.java b/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/output/SortingRFileOutputFormat.java
index 1fa8fdc..c54db06 100644
--- a/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/output/SortingRFileOutputFormat.java
+++ b/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/output/SortingRFileOutputFormat.java
@@ -105,17 +105,10 @@ public class SortingRFileOutputFormat extends OutputFormat<Text,Mutation> {
     
     // grab the configuration
     final Configuration conf = attempt.getConfiguration();
-    // create a filename
-    final String filenamePrefix = getPathName(conf);
-    final String taskID = attempt.getTaskAttemptID().toString();
     // grab the max size
     final long maxSize = getMaxBufferSize(conf);
-    // grab the FileSystem
-    final FileSystem fs = FileSystem.get(conf);
-    // create a default AccumuloConfiguration
-    final AccumuloConfiguration acuconf = AccumuloConfiguration.getDefaultConfiguration();
     
-    return new BufferingRFileRecordWriter(maxSize, acuconf, conf, filenamePrefix, taskID, fs);
+    return new BufferingRFileRecordWriter(maxSize, conf);
   }
   
 }

http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/e32632ca/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/protobuf/TermWeight.java
----------------------------------------------------------------------
diff --git a/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/protobuf/TermWeight.java b/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/protobuf/TermWeight.java
index bf5133f..7ae7aca 100644
--- a/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/protobuf/TermWeight.java
+++ b/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/protobuf/TermWeight.java
@@ -1,19 +1,3 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one or more
- * contributor license agreements.  See the NOTICE file distributed with
- * this work for additional information regarding copyright ownership.
- * The ASF licenses this file to You under the Apache License, Version 2.0
- * (the "License"); you may not use this file except in compliance with
- * the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
 // Generated by the protocol buffer compiler.  DO NOT EDIT!
 // source: TermWeight.proto
 
@@ -21,98 +5,239 @@ package org.apache.accumulo.examples.wikisearch.protobuf;
 
 public final class TermWeight {
   private TermWeight() {}
-  
-  public static void registerAllExtensions(com.google.protobuf.ExtensionRegistry registry) {}
-  
-  public static final class Info extends com.google.protobuf.GeneratedMessage {
+  public static void registerAllExtensions(
+      com.google.protobuf.ExtensionRegistry registry) {
+  }
+  public interface InfoOrBuilder
+      extends com.google.protobuf.MessageOrBuilder {
+
+    // required float normalizedTermFrequency = 1;
+    /**
+     * <code>required float normalizedTermFrequency = 1;</code>
+     */
+    boolean hasNormalizedTermFrequency();
+    /**
+     * <code>required float normalizedTermFrequency = 1;</code>
+     */
+    float getNormalizedTermFrequency();
+
+    // repeated uint32 wordOffset = 2;
+    /**
+     * <code>repeated uint32 wordOffset = 2;</code>
+     */
+    java.util.List<java.lang.Integer> getWordOffsetList();
+    /**
+     * <code>repeated uint32 wordOffset = 2;</code>
+     */
+    int getWordOffsetCount();
+    /**
+     * <code>repeated uint32 wordOffset = 2;</code>
+     */
+    int getWordOffset(int index);
+  }
+  /**
+   * Protobuf type {@code org.apache.accumulo.examples.wikisearch.protobuf.Info}
+   */
+  public static final class Info extends
+      com.google.protobuf.GeneratedMessage
+      implements InfoOrBuilder {
     // Use Info.newBuilder() to construct.
-    private Info() {
-      initFields();
+    private Info(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+      super(builder);
+      this.unknownFields = builder.getUnknownFields();
     }
-    
-    private Info(boolean noInit) {}
-    
+    private Info(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
     private static final Info defaultInstance;
-    
     public static Info getDefaultInstance() {
       return defaultInstance;
     }
-    
+
     public Info getDefaultInstanceForType() {
       return defaultInstance;
     }
-    
-    public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() {
-      return org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.internal_static_protobuf_Info_descriptor;
+
+    private final com.google.protobuf.UnknownFieldSet unknownFields;
+    @java.lang.Override
+    public final com.google.protobuf.UnknownFieldSet
+        getUnknownFields() {
+      return this.unknownFields;
+    }
+    private Info(
+        com.google.protobuf.CodedInputStream input,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      initFields();
+      int mutable_bitField0_ = 0;
+      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+          com.google.protobuf.UnknownFieldSet.newBuilder();
+      try {
+        boolean done = false;
+        while (!done) {
+          int tag = input.readTag();
+          switch (tag) {
+            case 0:
+              done = true;
+              break;
+            default: {
+              if (!parseUnknownField(input, unknownFields,
+                                     extensionRegistry, tag)) {
+                done = true;
+              }
+              break;
+            }
+            case 13: {
+              bitField0_ |= 0x00000001;
+              normalizedTermFrequency_ = input.readFloat();
+              break;
+            }
+            case 16: {
+              if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
+                wordOffset_ = new java.util.ArrayList<java.lang.Integer>();
+                mutable_bitField0_ |= 0x00000002;
+              }
+              wordOffset_.add(input.readUInt32());
+              break;
+            }
+            case 18: {
+              int length = input.readRawVarint32();
+              int limit = input.pushLimit(length);
+              if (!((mutable_bitField0_ & 0x00000002) == 0x00000002) && input.getBytesUntilLimit() > 0) {
+                wordOffset_ = new java.util.ArrayList<java.lang.Integer>();
+                mutable_bitField0_ |= 0x00000002;
+              }
+              while (input.getBytesUntilLimit() > 0) {
+                wordOffset_.add(input.readUInt32());
+              }
+              input.popLimit(limit);
+              break;
+            }
+          }
+        }
+      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+        throw e.setUnfinishedMessage(this);
+      } catch (java.io.IOException e) {
+        throw new com.google.protobuf.InvalidProtocolBufferException(
+            e.getMessage()).setUnfinishedMessage(this);
+      } finally {
+        if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
+          wordOffset_ = java.util.Collections.unmodifiableList(wordOffset_);
+        }
+        this.unknownFields = unknownFields.build();
+        makeExtensionsImmutable();
+      }
+    }
+    public static final com.google.protobuf.Descriptors.Descriptor
+        getDescriptor() {
+      return org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.internal_static_org_apache_accumulo_examples_wikisearch_protobuf_Info_descriptor;
     }
-    
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() {
-      return org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.internal_static_protobuf_Info_fieldAccessorTable;
+
+    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+        internalGetFieldAccessorTable() {
+      return org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.internal_static_org_apache_accumulo_examples_wikisearch_protobuf_Info_fieldAccessorTable
+          .ensureFieldAccessorsInitialized(
+              org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info.class, org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info.Builder.class);
     }
-    
+
+    public static com.google.protobuf.Parser<Info> PARSER =
+        new com.google.protobuf.AbstractParser<Info>() {
+      public Info parsePartialFrom(
+          com.google.protobuf.CodedInputStream input,
+          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+          throws com.google.protobuf.InvalidProtocolBufferException {
+        return new Info(input, extensionRegistry);
+      }
+    };
+
+    @java.lang.Override
+    public com.google.protobuf.Parser<Info> getParserForType() {
+      return PARSER;
+    }
+
+    private int bitField0_;
     // required float normalizedTermFrequency = 1;
     public static final int NORMALIZEDTERMFREQUENCY_FIELD_NUMBER = 1;
-    private boolean hasNormalizedTermFrequency;
-    private float normalizedTermFrequency_ = 0F;
-    
+    private float normalizedTermFrequency_;
+    /**
+     * <code>required float normalizedTermFrequency = 1;</code>
+     */
     public boolean hasNormalizedTermFrequency() {
-      return hasNormalizedTermFrequency;
+      return ((bitField0_ & 0x00000001) == 0x00000001);
     }
-    
+    /**
+     * <code>required float normalizedTermFrequency = 1;</code>
+     */
     public float getNormalizedTermFrequency() {
       return normalizedTermFrequency_;
     }
-    
+
     // repeated uint32 wordOffset = 2;
     public static final int WORDOFFSET_FIELD_NUMBER = 2;
-    private java.util.List<java.lang.Integer> wordOffset_ = java.util.Collections.emptyList();
-    
-    public java.util.List<java.lang.Integer> getWordOffsetList() {
+    private java.util.List<java.lang.Integer> wordOffset_;
+    /**
+     * <code>repeated uint32 wordOffset = 2;</code>
+     */
+    public java.util.List<java.lang.Integer>
+        getWordOffsetList() {
       return wordOffset_;
     }
-    
+    /**
+     * <code>repeated uint32 wordOffset = 2;</code>
+     */
     public int getWordOffsetCount() {
       return wordOffset_.size();
     }
-    
+    /**
+     * <code>repeated uint32 wordOffset = 2;</code>
+     */
     public int getWordOffset(int index) {
       return wordOffset_.get(index);
     }
-    
-    private void initFields() {}
-    
+
+    private void initFields() {
+      normalizedTermFrequency_ = 0F;
+      wordOffset_ = java.util.Collections.emptyList();
+    }
+    private byte memoizedIsInitialized = -1;
     public final boolean isInitialized() {
-      if (!hasNormalizedTermFrequency)
+      byte isInitialized = memoizedIsInitialized;
+      if (isInitialized != -1) return isInitialized == 1;
+
+      if (!hasNormalizedTermFrequency()) {
+        memoizedIsInitialized = 0;
         return false;
+      }
+      memoizedIsInitialized = 1;
       return true;
     }
-    
-    public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException {
+
+    public void writeTo(com.google.protobuf.CodedOutputStream output)
+                        throws java.io.IOException {
       getSerializedSize();
-      if (hasNormalizedTermFrequency()) {
-        output.writeFloat(1, getNormalizedTermFrequency());
+      if (((bitField0_ & 0x00000001) == 0x00000001)) {
+        output.writeFloat(1, normalizedTermFrequency_);
       }
-      for (int element : getWordOffsetList()) {
-        output.writeUInt32(2, element);
+      for (int i = 0; i < wordOffset_.size(); i++) {
+        output.writeUInt32(2, wordOffset_.get(i));
       }
       getUnknownFields().writeTo(output);
     }
-    
+
     private int memoizedSerializedSize = -1;
-    
     public int getSerializedSize() {
       int size = memoizedSerializedSize;
-      if (size != -1)
-        return size;
-      
+      if (size != -1) return size;
+
       size = 0;
-      if (hasNormalizedTermFrequency()) {
-        size += com.google.protobuf.CodedOutputStream.computeFloatSize(1, getNormalizedTermFrequency());
+      if (((bitField0_ & 0x00000001) == 0x00000001)) {
+        size += com.google.protobuf.CodedOutputStream
+          .computeFloatSize(1, normalizedTermFrequency_);
       }
       {
         int dataSize = 0;
-        for (int element : getWordOffsetList()) {
-          dataSize += com.google.protobuf.CodedOutputStream.computeUInt32SizeNoTag(element);
+        for (int i = 0; i < wordOffset_.size(); i++) {
+          dataSize += com.google.protobuf.CodedOutputStream
+            .computeUInt32SizeNoTag(wordOffset_.get(i));
         }
         size += dataSize;
         size += 1 * getWordOffsetList().size();
@@ -121,304 +246,368 @@ public final class TermWeight {
       memoizedSerializedSize = size;
       return size;
     }
-    
-    public static org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info parseFrom(com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data).buildParsed();
+
+    private static final long serialVersionUID = 0L;
+    @java.lang.Override
+    protected java.lang.Object writeReplace()
+        throws java.io.ObjectStreamException {
+      return super.writeReplace();
     }
-    
-    public static org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info parseFrom(com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+
+    public static org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info parseFrom(
+        com.google.protobuf.ByteString data)
         throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data, extensionRegistry).buildParsed();
+      return PARSER.parseFrom(data);
     }
-    
-    public static org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data).buildParsed();
+    public static org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info parseFrom(
+        com.google.protobuf.ByteString data,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data, extensionRegistry);
     }
-    
-    public static org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info parseFrom(byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+    public static org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info parseFrom(byte[] data)
         throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data, extensionRegistry).buildParsed();
+      return PARSER.parseFrom(data);
     }
-    
-    public static org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info parseFrom(java.io.InputStream input) throws java.io.IOException {
-      return newBuilder().mergeFrom(input).buildParsed();
+    public static org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info parseFrom(
+        byte[] data,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data, extensionRegistry);
     }
-    
-    public static org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info parseFrom(java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+    public static org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info parseFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return newBuilder().mergeFrom(input, extensionRegistry).buildParsed();
-    }
-    
-    public static org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException {
-      Builder builder = newBuilder();
-      if (builder.mergeDelimitedFrom(input)) {
-        return builder.buildParsed();
-      } else {
-        return null;
-      }
+      return PARSER.parseFrom(input);
     }
-    
-    public static org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info parseDelimitedFrom(java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+    public static org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info parseFrom(
+        java.io.InputStream input,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      Builder builder = newBuilder();
-      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
-        return builder.buildParsed();
-      } else {
-        return null;
-      }
+      return PARSER.parseFrom(input, extensionRegistry);
     }
-    
-    public static org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info parseFrom(com.google.protobuf.CodedInputStream input) throws java.io.IOException {
-      return newBuilder().mergeFrom(input).buildParsed();
+    public static org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info parseDelimitedFrom(java.io.InputStream input)
+        throws java.io.IOException {
+      return PARSER.parseDelimitedFrom(input);
     }
-    
-    public static org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info parseFrom(com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+    public static org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info parseDelimitedFrom(
+        java.io.InputStream input,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return newBuilder().mergeFrom(input, extensionRegistry).buildParsed();
+      return PARSER.parseDelimitedFrom(input, extensionRegistry);
     }
-    
-    public static Builder newBuilder() {
-      return Builder.create();
+    public static org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info parseFrom(
+        com.google.protobuf.CodedInputStream input)
+        throws java.io.IOException {
+      return PARSER.parseFrom(input);
     }
-    
-    public Builder newBuilderForType() {
-      return newBuilder();
+    public static org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info parseFrom(
+        com.google.protobuf.CodedInputStream input,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws java.io.IOException {
+      return PARSER.parseFrom(input, extensionRegistry);
     }
-    
+
+    public static Builder newBuilder() { return Builder.create(); }
+    public Builder newBuilderForType() { return newBuilder(); }
     public static Builder newBuilder(org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info prototype) {
       return newBuilder().mergeFrom(prototype);
     }
-    
-    public Builder toBuilder() {
-      return newBuilder(this);
-    }
-    
-    public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> {
-      private org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info result;
-      
-      // Construct using protobuf.TermWeight.Info.newBuilder()
-      private Builder() {}
-      
-      private static Builder create() {
-        Builder builder = new Builder();
-        builder.result = new org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info();
-        return builder;
+    public Builder toBuilder() { return newBuilder(this); }
+
+    @java.lang.Override
+    protected Builder newBuilderForType(
+        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+      Builder builder = new Builder(parent);
+      return builder;
+    }
+    /**
+     * Protobuf type {@code org.apache.accumulo.examples.wikisearch.protobuf.Info}
+     */
+    public static final class Builder extends
+        com.google.protobuf.GeneratedMessage.Builder<Builder>
+       implements org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.InfoOrBuilder {
+      public static final com.google.protobuf.Descriptors.Descriptor
+          getDescriptor() {
+        return org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.internal_static_org_apache_accumulo_examples_wikisearch_protobuf_Info_descriptor;
       }
-      
-      protected org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info internalGetResult() {
-        return result;
+
+      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+          internalGetFieldAccessorTable() {
+        return org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.internal_static_org_apache_accumulo_examples_wikisearch_protobuf_Info_fieldAccessorTable
+            .ensureFieldAccessorsInitialized(
+                org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info.class, org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info.Builder.class);
       }
-      
-      public Builder clear() {
-        if (result == null) {
-          throw new IllegalStateException("Cannot call clear() after build().");
+
+      // Construct using org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info.newBuilder()
+      private Builder() {
+        maybeForceBuilderInitialization();
+      }
+
+      private Builder(
+          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+        super(parent);
+        maybeForceBuilderInitialization();
+      }
+      private void maybeForceBuilderInitialization() {
+        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
         }
-        result = new org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info();
+      }
+      private static Builder create() {
+        return new Builder();
+      }
+
+      public Builder clear() {
+        super.clear();
+        normalizedTermFrequency_ = 0F;
+        bitField0_ = (bitField0_ & ~0x00000001);
+        wordOffset_ = java.util.Collections.emptyList();
+        bitField0_ = (bitField0_ & ~0x00000002);
         return this;
       }
-      
+
       public Builder clone() {
-        return create().mergeFrom(result);
+        return create().mergeFrom(buildPartial());
       }
-      
-      public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() {
-        return org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info.getDescriptor();
+
+      public com.google.protobuf.Descriptors.Descriptor
+          getDescriptorForType() {
+        return org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.internal_static_org_apache_accumulo_examples_wikisearch_protobuf_Info_descriptor;
       }
-      
+
       public org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info getDefaultInstanceForType() {
         return org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info.getDefaultInstance();
       }
-      
-      public boolean isInitialized() {
-        return result.isInitialized();
-      }
-      
+
       public org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info build() {
-        if (result != null && !isInitialized()) {
+        org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info result = buildPartial();
+        if (!result.isInitialized()) {
           throw newUninitializedMessageException(result);
         }
-        return buildPartial();
-      }
-      
-      private org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info buildParsed() throws com.google.protobuf.InvalidProtocolBufferException {
-        if (!isInitialized()) {
-          throw newUninitializedMessageException(result).asInvalidProtocolBufferException();
-        }
-        return buildPartial();
+        return result;
       }
-      
+
       public org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info buildPartial() {
-        if (result == null) {
-          throw new IllegalStateException("build() has already been called on this Builder.");
+        org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info result = new org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info(this);
+        int from_bitField0_ = bitField0_;
+        int to_bitField0_ = 0;
+        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+          to_bitField0_ |= 0x00000001;
         }
-        if (result.wordOffset_ != java.util.Collections.EMPTY_LIST) {
-          result.wordOffset_ = java.util.Collections.unmodifiableList(result.wordOffset_);
+        result.normalizedTermFrequency_ = normalizedTermFrequency_;
+        if (((bitField0_ & 0x00000002) == 0x00000002)) {
+          wordOffset_ = java.util.Collections.unmodifiableList(wordOffset_);
+          bitField0_ = (bitField0_ & ~0x00000002);
         }
-        org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info returnMe = result;
-        result = null;
-        return returnMe;
+        result.wordOffset_ = wordOffset_;
+        result.bitField0_ = to_bitField0_;
+        onBuilt();
+        return result;
       }
-      
+
       public Builder mergeFrom(com.google.protobuf.Message other) {
         if (other instanceof org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info) {
-          return mergeFrom((org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info) other);
+          return mergeFrom((org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info)other);
         } else {
           super.mergeFrom(other);
           return this;
         }
       }
-      
+
       public Builder mergeFrom(org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info other) {
-        if (other == org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info.getDefaultInstance())
-          return this;
+        if (other == org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info.getDefaultInstance()) return this;
         if (other.hasNormalizedTermFrequency()) {
           setNormalizedTermFrequency(other.getNormalizedTermFrequency());
         }
         if (!other.wordOffset_.isEmpty()) {
-          if (result.wordOffset_.isEmpty()) {
-            result.wordOffset_ = new java.util.ArrayList<java.lang.Integer>();
+          if (wordOffset_.isEmpty()) {
+            wordOffset_ = other.wordOffset_;
+            bitField0_ = (bitField0_ & ~0x00000002);
+          } else {
+            ensureWordOffsetIsMutable();
+            wordOffset_.addAll(other.wordOffset_);
           }
-          result.wordOffset_.addAll(other.wordOffset_);
+          onChanged();
         }
         this.mergeUnknownFields(other.getUnknownFields());
         return this;
       }
-      
-      public Builder mergeFrom(com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+
+      public final boolean isInitialized() {
+        if (!hasNormalizedTermFrequency()) {
+          
+          return false;
+        }
+        return true;
+      }
+
+      public Builder mergeFrom(
+          com.google.protobuf.CodedInputStream input,
+          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
           throws java.io.IOException {
-        com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(this.getUnknownFields());
-        while (true) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              this.setUnknownFields(unknownFields.build());
-              return this;
-            default: {
-              if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) {
-                this.setUnknownFields(unknownFields.build());
-                return this;
-              }
-              break;
-            }
-            case 13: {
-              setNormalizedTermFrequency(input.readFloat());
-              break;
-            }
-            case 16: {
-              addWordOffset(input.readUInt32());
-              break;
-            }
-            case 18: {
-              int length = input.readRawVarint32();
-              int limit = input.pushLimit(length);
-              while (input.getBytesUntilLimit() > 0) {
-                addWordOffset(input.readUInt32());
-              }
-              input.popLimit(limit);
-              break;
-            }
+        org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info parsedMessage = null;
+        try {
+          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+          parsedMessage = (org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info) e.getUnfinishedMessage();
+          throw e;
+        } finally {
+          if (parsedMessage != null) {
+            mergeFrom(parsedMessage);
           }
         }
+        return this;
       }
-      
+      private int bitField0_;
+
       // required float normalizedTermFrequency = 1;
+      private float normalizedTermFrequency_ ;
+      /**
+       * <code>required float normalizedTermFrequency = 1;</code>
+       */
       public boolean hasNormalizedTermFrequency() {
-        return result.hasNormalizedTermFrequency();
+        return ((bitField0_ & 0x00000001) == 0x00000001);
       }
-      
+      /**
+       * <code>required float normalizedTermFrequency = 1;</code>
+       */
       public float getNormalizedTermFrequency() {
-        return result.getNormalizedTermFrequency();
+        return normalizedTermFrequency_;
       }
-      
+      /**
+       * <code>required float normalizedTermFrequency = 1;</code>
+       */
       public Builder setNormalizedTermFrequency(float value) {
-        result.hasNormalizedTermFrequency = true;
-        result.normalizedTermFrequency_ = value;
+        bitField0_ |= 0x00000001;
+        normalizedTermFrequency_ = value;
+        onChanged();
         return this;
       }
-      
+      /**
+       * <code>required float normalizedTermFrequency = 1;</code>
+       */
       public Builder clearNormalizedTermFrequency() {
-        result.hasNormalizedTermFrequency = false;
-        result.normalizedTermFrequency_ = 0F;
+        bitField0_ = (bitField0_ & ~0x00000001);
+        normalizedTermFrequency_ = 0F;
+        onChanged();
         return this;
       }
-      
+
       // repeated uint32 wordOffset = 2;
-      public java.util.List<java.lang.Integer> getWordOffsetList() {
-        return java.util.Collections.unmodifiableList(result.wordOffset_);
+      private java.util.List<java.lang.Integer> wordOffset_ = java.util.Collections.emptyList();
+      private void ensureWordOffsetIsMutable() {
+        if (!((bitField0_ & 0x00000002) == 0x00000002)) {
+          wordOffset_ = new java.util.ArrayList<java.lang.Integer>(wordOffset_);
+          bitField0_ |= 0x00000002;
+         }
+      }
+      /**
+       * <code>repeated uint32 wordOffset = 2;</code>
+       */
+      public java.util.List<java.lang.Integer>
+          getWordOffsetList() {
+        return java.util.Collections.unmodifiableList(wordOffset_);
       }
-      
+      /**
+       * <code>repeated uint32 wordOffset = 2;</code>
+       */
       public int getWordOffsetCount() {
-        return result.getWordOffsetCount();
+        return wordOffset_.size();
       }
-      
+      /**
+       * <code>repeated uint32 wordOffset = 2;</code>
+       */
       public int getWordOffset(int index) {
-        return result.getWordOffset(index);
+        return wordOffset_.get(index);
       }
-      
-      public Builder setWordOffset(int index, int value) {
-        result.wordOffset_.set(index, value);
+      /**
+       * <code>repeated uint32 wordOffset = 2;</code>
+       */
+      public Builder setWordOffset(
+          int index, int value) {
+        ensureWordOffsetIsMutable();
+        wordOffset_.set(index, value);
+        onChanged();
         return this;
       }
-      
+      /**
+       * <code>repeated uint32 wordOffset = 2;</code>
+       */
       public Builder addWordOffset(int value) {
-        if (result.wordOffset_.isEmpty()) {
-          result.wordOffset_ = new java.util.ArrayList<java.lang.Integer>();
-        }
-        result.wordOffset_.add(value);
+        ensureWordOffsetIsMutable();
+        wordOffset_.add(value);
+        onChanged();
         return this;
       }
-      
-      public Builder addAllWordOffset(java.lang.Iterable<? extends java.lang.Integer> values) {
-        if (result.wordOffset_.isEmpty()) {
-          result.wordOffset_ = new java.util.ArrayList<java.lang.Integer>();
-        }
-        super.addAll(values, result.wordOffset_);
+      /**
+       * <code>repeated uint32 wordOffset = 2;</code>
+       */
+      public Builder addAllWordOffset(
+          java.lang.Iterable<? extends java.lang.Integer> values) {
+        ensureWordOffsetIsMutable();
+        super.addAll(values, wordOffset_);
+        onChanged();
         return this;
       }
-      
+      /**
+       * <code>repeated uint32 wordOffset = 2;</code>
+       */
       public Builder clearWordOffset() {
-        result.wordOffset_ = java.util.Collections.emptyList();
+        wordOffset_ = java.util.Collections.emptyList();
+        bitField0_ = (bitField0_ & ~0x00000002);
+        onChanged();
         return this;
       }
-      
-      // @@protoc_insertion_point(builder_scope:protobuf.Info)
+
+      // @@protoc_insertion_point(builder_scope:org.apache.accumulo.examples.wikisearch.protobuf.Info)
     }
-    
+
     static {
       defaultInstance = new Info(true);
-      org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.internalForceInit();
       defaultInstance.initFields();
     }
-    
-    // @@protoc_insertion_point(class_scope:protobuf.Info)
+
+    // @@protoc_insertion_point(class_scope:org.apache.accumulo.examples.wikisearch.protobuf.Info)
   }
-  
-  private static com.google.protobuf.Descriptors.Descriptor internal_static_protobuf_Info_descriptor;
-  private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_protobuf_Info_fieldAccessorTable;
-  
-  public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() {
+
+  private static com.google.protobuf.Descriptors.Descriptor
+    internal_static_org_apache_accumulo_examples_wikisearch_protobuf_Info_descriptor;
+  private static
+    com.google.protobuf.GeneratedMessage.FieldAccessorTable
+      internal_static_org_apache_accumulo_examples_wikisearch_protobuf_Info_fieldAccessorTable;
+
+  public static com.google.protobuf.Descriptors.FileDescriptor
+      getDescriptor() {
     return descriptor;
   }
-  
-  private static com.google.protobuf.Descriptors.FileDescriptor descriptor;
+  private static com.google.protobuf.Descriptors.FileDescriptor
+      descriptor;
   static {
-    java.lang.String[] descriptorData = {"\n\020TermWeight.proto\022\010protobuf\";\n\004Info\022\037\n\027"
-        + "normalizedTermFrequency\030\001 \002(\002\022\022\n\nwordOff" + "set\030\002 \003(\rB\014\n\010protobufH\001"};
-    com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
-      public com.google.protobuf.ExtensionRegistry assignDescriptors(com.google.protobuf.Descriptors.FileDescriptor root) {
-        descriptor = root;
-        internal_static_protobuf_Info_descriptor = getDescriptor().getMessageTypes().get(0);
-        internal_static_protobuf_Info_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable(
-            internal_static_protobuf_Info_descriptor, new java.lang.String[] {"NormalizedTermFrequency", "WordOffset",}, org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info.class,
-            org.apache.accumulo.examples.wikisearch.protobuf.TermWeight.Info.Builder.class);
-        return null;
-      }
+    java.lang.String[] descriptorData = {
+      "\n\020TermWeight.proto\0220org.apache.accumulo." +
+      "examples.wikisearch.protobuf\";\n\004Info\022\037\n\027" +
+      "normalizedTermFrequency\030\001 \002(\002\022\022\n\nwordOff" +
+      "set\030\002 \003(\rB4\n0org.apache.accumulo.example" +
+      "s.wikisearch.protobufH\001"
     };
-    com.google.protobuf.Descriptors.FileDescriptor.internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] {},
-        assigner);
+    com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
+      new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
+        public com.google.protobuf.ExtensionRegistry assignDescriptors(
+            com.google.protobuf.Descriptors.FileDescriptor root) {
+          descriptor = root;
+          internal_static_org_apache_accumulo_examples_wikisearch_protobuf_Info_descriptor =
+            getDescriptor().getMessageTypes().get(0);
+          internal_static_org_apache_accumulo_examples_wikisearch_protobuf_Info_fieldAccessorTable = new
+            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
+              internal_static_org_apache_accumulo_examples_wikisearch_protobuf_Info_descriptor,
+              new java.lang.String[] { "NormalizedTermFrequency", "WordOffset", });
+          return null;
+        }
+      };
+    com.google.protobuf.Descriptors.FileDescriptor
+      .internalBuildGeneratedFileFrom(descriptorData,
+        new com.google.protobuf.Descriptors.FileDescriptor[] {
+        }, assigner);
   }
-  
-  public static void internalForceInit() {}
-  
+
   // @@protoc_insertion_point(outer_class_scope)
 }